Earlier this month I promised to show you a simple example of regularization. Well, here it is. This is a particular combination of integrals I've been working with a lot lately:
A quick look at the formulas shows you that both integrands have singularities at and .
OK, well, that's why we have two integrands. We can change variables from , subtract them, and the singularities will cancel out, right? You can do this integral by hand in polar coordinates, or just pop it into Mathematica:
Just one problem: that's not the right answer! You can tell because the value of this integral had better depend on , but this expression doesn't. This isn't anything so mundane as a bug in Mathematica. It's a "bug" in math itself.
See, when you have an undefined quantity, like a divergent integral, you can't just keep doing math with the expression that represents it. That can lead to nonsensical results, like this classic invalid proof that :
Let and . Since , then and , so
Dividing both sides by , we get
and since , that gives or . But we know that , therefore
In this fake proof, it's pretty easy to find the bad step: dividing by is dividing by zero, and of course you can't divide by zero. With the divergent integrals, something similar is going on, but it's a bit trickier to figure out where the math goes wrong. The bad step is changing variables at , because, in a sense, you wind up losing your new boundary condition into the singularity.
Fortunately, it's not too hard to come up with a way to get around this. If is the problematic point, just leave it out! More specifically, you can cut out a small region around that singularity, do the integral over the rest of the plane, and then see if it's well-defined in the limit as the excised region shrinks to zero. This is called cutoff regularization, and it's the simplest of several techniques that you can use to force sensible results out of combinations of divergent integrals.
So instead of integrating over the entire plane, we're integrating over or for some value that is the radius of the "hole" around the singularity.
Now we can change variables to in the second integral and keep track of what happens at the boundary of the integration region. (There's also an outer boundary at to worry about, but the function is zero there so that one actually doesn't matter. But you could use cutoff regularization on that as well.)
We'll have to correct for the difference between the two regions of integration. Suppose , then you can split the second integral into two pieces: one over and one over .
The first of these is the term you'd miss by taking the naive approach I described at the beginning.
Finally, it's safe to combine the first two integrals, getting the same result as before. And in the last term, we can use the series expansion of the integrand at to do it by hand. I'll skip the details, though, and just offer up the answer:
This shows the dependence on , as it should. So this expression is a nice, sensible answer.
If you were really thinking, you might have wondered why we had to impose the same cutoff, , on both integrals. After all, they're over different variables; why should their boundaries necessarily have any relation to each other? Well, that's just the way cutoff regularization works. You could just as well use a different regularization procedure where you cut off one integral at and the other at , and you'd get a different answer. Technically, all we've really done is take the arbitrariness out of the singularities and move it into the regularization procedure. That really doesn't matter, though; what's more important is that different calculations are done using the same method of regularization, so that they're all consistent with each other.