Analyzing Coupled Systems
Designing new optical systems from the ground up can be an intimidating task, especially if the system we are trying to make is one that we must get right before anything can be built because the critical components simply do not yet exist. It’s merely a figment of our imagination and it is actually part of our task to define it. Similar situations will arise when we explore different designs and want to learn how things actually work before taking on the task of actually building it.
So let’s picture this. You need to find the requirements for a system that we’ve never studied before and if there is a scientific paper that maybe is related to what we need, are we even going to find it? It’s not uncommon that we reinvent the wheel. Dirac reinvented the Clifford algebra when he derived the quantum equations for the electron, so obviously this happens to the best of us. So let’s just assume that anyway we slice it, the buck stops with us.
The go-to strategy at this point would be to get a model going—a map of sorts, but not just any map; it needs to be detailed enough to navigate through all the essential twists and turns (read: degrees of freedom), yet not so intricate that we don’t see the woods for the trees. Here lies our first challenge: unless this machine is a close cousin of something we've already mastered, our map is more guesswork than gospel. We're somewhat in the dark about which paths (degrees of freedom) are the relevant ones, and which are just details.
Without the right compass (tools) to navigate the complexity, we're at risk of oversimplifying our map, constrained more by our imagination than by the actual terrain.
Reducing Complexity
One way to deal with this, that I have found to work well is to employ a well-known mathematical expansion around the ideal state of the system being studied, essentially a Maclaurin expansion. Since the tool we are looking for is one frees our analysis from our own mental limitations, it must be one that allows us to probe our system with a large number of independent variables.
There are different ways to approach this, but one that I have found to work very well is the so called Latin Hypercube Sampling (LHS), sometimes also referred to as a space-filling design.
Generating a LHS takes some time, but in most cases its usually less than a few minutes, even if the number of independent variables is large. The time to generating a LHS grows linearly with the number of variables and to the 3rd power with the number of samples. Luckily, this is only polynomial complexity, or "easy” if you ask Computational Complexity people.
An example of where a large number of DOF may be interesting is when one wants to derive a specification for an optical system and the DOF are the Zernike polynomial coefficients.
What the LHS gives us is a set of uniformly distributed samples over a multi-dimensional space, but it is still up to us to ask the right question, and in this case, there are practically no wrong questions.
So, for example, using this set of samples, one can set up a set of simulations that probes the variance of property, or we can simulate how some deterministic property depends on our DOF. Both can be very valuable, but before we dive deeper into that, let’s talk about how to get some results to analyze.
Second order exapansion
As simple as this may sound, this is a very valuable approach that offers insights that quite often turn into aha moments.
From the LHS we have a set of points, and using these sampling points into our multi-dimensional problem, we run an equal amount of simulations to get a numerical result for each. However, having done all that is not very enlighting. We need to reduce this complexity. We need to summarize all these simulations, which could easily be several thousands, into something we can grasp, something that sorts the chaff from the wheat so to speak.
The way to do this is to try to express our results using a second-order expansion,
This is very similar to the way we would express a MacLaurin expansion but the point is that the gradient and the Hessian are still unknown and we need to find them. Can we do that?
Yes, we can. We have plenty of information and finding the coeffieient of v and H is a linear problem. For a 2 DOF, we have 1 unknown “c”, to unknowns in the linear term “v” and 2 + 1 in “H” because H is symmetric. For 3 DOF, we have 1 + 3 + 3 + 2 + 1 and so on. This is just linear algebra. Our thousands of simulations and the More Penrose pseudo inverse will give us those unknowns quicker than we can say “numerical solution”.
To the good stuff
And it is now that the interesting stuff begins. Assuming that our 2nd order expansions is good, we have reduced often thousands of simulations into a model that we now can begin to take apart even further.
The first thing one usually would do after getting the expansion is to check how well does this expansion reproduces the thousands of simulated results. Quite often, the correlation between the model and the simulations is larger than 99%. I would consider 95% as barely useful but the usual result is around 99% or better. It will of course depend on what the underlying prorperty looks like, if our numerical model is for a statistical property or not.
Before we have played with this model, we don’t know that but I have yet to find a case that couldn’t be transformed in some way that allow this expansion to capture essentially all variance from our simulations.
Analyzing the expansion
Now we come to the part where we can reap the benefits of our hard work. There are typically three types of results, either the expansion is dominated by the linear term, or the square term or it is mixed, which actually is quite interesting because it tells us that our expansion is not around a minimum.
If the linear term dominates, well, I guess that’s it. No interactions between the DOF to talk about. Case closed. Our system was simple, and if we didn’t get that before, we certainly do now.
It gets more interesting when the second order term dominates the expansion because now, there is more to learn. What to do here is definitely to call more linear algebra into service and rewrite H as,
using the Singular Value Decomposition, or SVD for short.
If you've come this far and choose not to seize this opportunity, it would be tantamount to Neo taking the blue pill, opting to sleep and believe whatever he wants to believe. Never take the blue pill.
If your system needs to be expressed in terms of a second order matrix, you need to stay in wonderland and let the SVD show you how deep the rabit hole goes.
This is not just a theoretical game of numbers. You may need to know this when designing your optics because, not only does it tell you that certain combinations of aberration need to be optimized together, the diagonal elements of also tell you how important each of those combinations really are which is an invaluable hint what the merit function should suppress the most.
So, what about the mixed case, where both the linear part and the square part contribute significantly to the model? This is actually a really interesting one because it tells us that, contrary to your initial belief, the system was not expanded around the ideal state.
This case need requires a little caution because the solution is model dependent and even then, it may point to a new minumum i outside the volume probed by the initial set of sampling points. It may still be OK but it cannot be taken for granted.
It’s example time
Suppose we would apply this theory to something like fiber coupling. Staying with the spirit that we don’t know how it works, we use the Zernike polynomials as our degrees of freedom and, as our test function, we compute the insertion loss as a function of our degrees of freedom.
First, we would define how many DOF we are interested in studying. Let’s say that we stop at Z16. We know Z1 usually doesn’t matter but those that follow do, so, 15 DOF.
Next, we generate a 1024 sample points over 15 DOF using the LHS. The LHS usually generates sample points over a hypercube so we need to scale those values to something reasonable. If we know nothing about the subject, then probably we should start small, really small. Then, maybe repeat the process until we find that the model we compute no longer fits the data very well or until we reach a range of values that we believe to be relevant to our system.
Now, we jump a little in time and 1024 simulations of insertion loss later, we have our function f(x,….) and something to model. (This simulation assumed an apodization of 2 and a gaussian fundamental mode. This blog is not so much about single-mode fibers so forgive the simplified assumptions about single-mode fibers.)
Once we find the coefficents of c, v, and H, we can have a look how the model captured our simulations,
As expected, we got a good correlation between the simulations and the model expansion. Since the physical optics is only described by the overlap integral, that was to be expected. After all , this is supposed to be an example to learn from.
Now, we can also have a look at how much of this correlation comes from the linear term,
And again as expected, the correlation is a (statistical) zero. We modelled something which has a minimum at zero aberrations. The linear part is expected to vanish. So now, lets have a look at the square term.
I hope that by now, it’s pretty obvious that we are going to use the Singular Value Decomposition, remember the red pill, right? So, we throw our H into the SVD and the first thing to scrutinize are the singular values,
The actual values on the Y-axis are not so intersting, after all, they depend on the volume of our test hypercube we used to sample our problem. What’s more interesting is the structure. Our problem has been reduced to almost 3 independent variables, plus 2 much less important ones. Let’s have a look,
Time to sum up.
We started with 1024 simulated values over 15 dimension of something we (pretended to) know nothing about and ended up with 3 major contributions to a property we wanted to understand.
This is not an unusual result when doing this type of analysis. Perhaps the correlation is not quite as good as in this constructed example, but with some transformation of the data, the 2nd order model can capture essentially all the variance of the simulation.
Perhaps the most valuable aspect of this type of analysis is that we can take on a much larger number of degrees of freedom and not restrict ourselves by our ability to untangle the complexity, or to connect to the map metaphor, there are no trees. Woods and hills is all we see. From this vantage point, the big picture is all there is, and all we paid for it was computer time.
This blog dives deep into the philosophy of Senslogic and what we are trying to accomplish accomplish, and how the WaveMe toolbox fits into this picture.