Saturday, February 19, 2022

The Mind of God

Not that anybody reads the blog, but I guess I should post that I moved by blog to The Mind of God. It's been that way for a while, actually. Anyway, the stuff I have written there is way cooler than the stuff currently on here.

Tuesday, June 25, 2019

YARP (Yet Another Rant about Programming)

Imagine that, as a kid, you really wanted to learn an instrument.  Eventually, after lots of nagging, and some luck, your parents get you a cheap $50 guitar.

Excited, you strum the guitar, imagining that you will hear the overdriven birth cry of of some guitar-rock god, even though your guitar is acoustic, and you have no amplifier.  Instead, you hear shit.  And your hand starts cramping.

But, no matter.  Undeterred, you convince your parents to buy you some books, or videos on how to play the guitar, and slowly, you learn the ropes.  While all your peers spend their time in sports, or playing video games, you spend yours honing your guitar skills, until one day you are so good, that your parents to get you a professional guitar, and amplifier.  Equipped with this new tool, you transcend your previous limitations, and you decide that you wish to dedicate the rest of your life to being a musician.

So, you go to music school, where, to round you out as a musician, they require you to learn piano, some wind instrument, music theory, and composition.  No sweat.  Yes, there are many differences between the skills that guitar requires, and what these other disciplines require, but much of what you have learned, while playing guitar transfers over.

Sure, you never formally learned what a key was, or any of the myriad scales, but years of guitar playing had prepared you enough so that it wasn't too difficult to learn these new rules.

Then, you graduate, and are immediately hired as the guitar player for a band.

That's when the trouble starts.

You see, in school, they didn't tell you about what makes a professional musician valuable, and, in fact, you are surprised to learn that it is not actually knowing how to play one of several instruments, or even knowing music theory.

No, what makes a musician valuable is staying up to date with the latest instruments.

In fact, every year, a new instrument that purports to be more versatile, cheaper, and have better sound quality, than the previous year's instrument, is released, and unless you keep up to date with these latest models, you quickly become outmoded by your peers.  In fact, most of your peers already know these technologies, and they're the ones given all the cool solos, while you are stuck as a backup, or working with sound levels on recordings, which, by the way, are required to be released every two weeks--more on that in a second.

That Les Paul you got in 2010 is so yesterday.  These days, everybody uses a Xerxes Ubaphone with an Apache Goliath RDSS to release their latest tracks.

You don't know what an Ubaphone is, or what a RDSS is, but you jump in, and try learning it.  After all, you have years of music school, and self-education under your belt.  How hard could it be?

Well...

So, you find out that a Xerxes Ubaphone is not actually an instrument in itself, but, rather a complex framework of different connectors, for different instrument parts.  This is why it's so flexible.  With Ubaphone, you can create virtually any instrument you want, which seems pretty cool.  Except for the fact that it is so complex that its manual is 500 pages.  On the other hand, it has a thriving online community of musicians, who use it.  In fact, it was created by musicians, and is open source.

Additionally, you find that RDSS stands for Remote Distributed Sound System.  This is not your father's sound system.  No, it is a scalable, high fidelity, sound system that enables you to layer over millions of tracks simultaneously, and broadcast it over a virtually unlimited number of speakers, with sophisticated nose filtering technologies, which eliminates sound clipping, feedback gain, and other such nuisances.  And these speakers can be anywhere in the world, in relation to each other.

Sounds pretty cool, right?

Well, it also has a manual of 500 pages.  On top of that, its documentation assumes that most of its users have experience with an Ubaphone, so that, while it can hook into a regular instrument, all of its examples require you to have a fully deployed Ubaphone of some kind working.

Oh, and that if you use Xerxes Ubaphone, it needs be be a 2018 version, because previous versions had a vulnerability that, despite the Goliath RDSS' claim to eliminate feedback gain, could lead to a severe feeback gain, which tended to blow out half of the speakers in the sound system.

You groan, but you're a good sport.  So, you set about reading the manuals, and start with the Ubaphone.  The first thing you find is that, when you first unpackage the Xerxes Ubaphone it does not come ready to use.

Nope!  You have to configure it, by setting a bunch of dials, and buttons, and to your dismay, even though what you're reading claims to be a manual, it fails to provide even one example of a basic configuration that you can use in most reasonable cases.  Instead, it tells you to consult a chapter on "Configuring for different backends."

Not sure what a backend is, you look up this chapter, only to be assaulted with a slew of different types of configurations for things you have never heard of, with names that fail to convey what they actually do:

Apache Vindaloo, Skrunk, GPI, Zogaz, and so on...

It turns out that you can't just learn Xerxes Ubaphone.  You have to be familiar with at least one of these technologies.

That's because a Xerxes Ubaphone does not actually handle transformations of sound signals between parts, but only offers an interface for doing so.  The thing that really transforms sound signals is called a backend, and is handled by one of the gadgets above.

Well, Apache Vindaloo, like all Apache Musical Foundation products is well supported with a thriving online community--and a 500 page manual.

At this point, since you have already wasted half a day on documentation, you don't wish to read yet another fucking manual.

So, you try to see if there is some quickstart guide to setting up Apache Vindaloo, which, heavens be praised, you find pretty quickly.

In fact, it is published by Apache, so if anything, it should be the authoritative source on setting up Xerxes Ubaphone with Apache Vindaloo quickly, right?  Right?

Well, so far, it seems that way.  It turns out that even though the manual didn't mention it, the Xerxes Ubaphone comes preconfigured to work with an Apache Vindaloo, and, in fact, the Vindaloo comes with all 2015 distributions, and later--and that was that mysterious thingamagig you found bubblewrapped, and unlabeled with your Ubaphone.  Even better, the quickstart guide tells you precisely the steps to follow to get a basic Xerxes Ubaphone with an Apache Vindaloo Backend up and working, and even throws in attaching a Goliath RDSS!

So, you follow it, and by the end of the day, you now have a fully configured system.  Excited, you press a key, AND IT DOESN'T FUCKING WORK.

It turns out that the quickstart guide was published in 2015, and is outdated.

This process continues for a week, until, finally, after days of talking with fellow musicians, and desperately scanning forums and google groups, you identify the problem: the Ubaphone requires a specific up-to-date power adapter, and apparently, this was such common knowledge to anybody who had ever used an Ubaphone, nobody ever bothered to mention it.  You find this out, after the lead musician berates you, asking you what kind of musician you are anyway, if you don't even know something so basic?

If you think this is the end of your woes, you are sorely mistaken, my friend.  You see, the other major thing about modern professional musicianship nobody fucking bothered to tell you is that musicians are treated like mules.  In fact, a methodology of producing music has taken the music industry by storm.  It is known as VERSATILE, and all managers, and producers love it, even though it is the scourge of the earth.  What distinguishes VERSATILE from previous ways of producing music, like improvisation, and traditional composition, is that it is a results-driven methodology built around the following mantra:

"Release musical tracks early, and often."

All the biggest labels follow this methodology, because it offers a reliable way to mass-produce music that everybody listens to.

To achieve this goal, VERSATILE prescribes the following practices, and guidelines:

  1. Split up your planning into 2-week long intervals called "sprints."
  2. Each sprint is expected to enrich the current track being produced with new features, or patch errors in the existing track.
  3. Frequently release of variations of the same music track.  It doesn't matter, if the track lacks a resolution, or that the vocalist stutters during the final verse, or that the 2nd verse is just the lyrical filler, "Boodabee, Babbadoo, Quabbawee, Shabbadoo", because they lyricist couldn't think of clever enough lyrics.  That can be fixed with the next sprint.  Just get a minimum viable track out there.
  4. Speaking of which, whenever developing a track, just focus on short-term, realizable goals.  In fact, focusing on the long term, and painstakingly working out a plan for making a masterpiece is frowned upon by management as emblematic of the old way of doing things, which, you used to call "giving a damn about your art", but which they call the CATARACT approach.  In fact, CATARACT is used by everybody in the music industry to derisively refer to any practice that doesn't conform with VERSATILE, even if it fits your workflow better than anything VERSATILE has to offer.


The worst part about VERSATILE is how it is abused to force musicians to work 60-80 hour workweeks.  You see, because you need a minimum viable track in 2-week sprints, the producer doesn't want to hear it if, say, your efforts of laying over your Ubaphone track configured for Les Paul Acoustic Guitar were hampered by the lack of support for Les Paul Guitar, even though he requested it.  You had better figure something out, and fast!  Better get yourself a pot of coffee, and say goodbye to coming home to your wife and kids, tonight!

If you are thinking to yourself, "No musician would ever want to become a professional musician under those ridiculous constraints," it is safe to say you are not alone.

On the other hand, the above scenario is precisely the kind of bullshit that professional programmers have to go through, every fucking day.

In fact, I was inspired to write this rant, as a result of attempting to deploy something called JanusGraph, and the process of learning it is not too far off from the preposterous Xerxes Ubaphone scenario I mentioned above.

Look, I don't expect programming to be easy.  Writing code that does nontrivial things correctly, efficiently, and works as expected in other machines but the one you wrote it on is difficult.  I accept that.

What I have difficulty accepting is the other bullshit--this whole process of spending 70% of my time understanding an ever changing suite of tools instead of actually using them.  These tools purport to somehow make my life better and increase my productivity, but they don't, because I can't actually do my job until I have mastered configuring them in the first place, and then I have to scramble in the little time I have left to create a minimum viable product so that my manager isn't shit on by his manager.

In fact, this is such a problem that I often prefer to write my own utility, rather than suffer the dependency, and configuration hell of learning a new library, or framework.  Don't chastise me for reinventing the wheel, when learning how to use any of the available ones requires me to expend weeks of mental energy just getting it to roll--weeks that I could have devoted to actually doing something interesting--programming!  What a novel concept!

Look, I know that documentation is hard, and interfaces evolve, but there's got to be a better way of doing this, because it is not the programming that I fell in love with.  Hell, it isn't even programming!

Wednesday, March 23, 2016

Keeping Track of Complicated Algebra

Lately, I have been doing a lot or independent work in classical electromagnetism.  I have found, as anyone who has ever derived the Poynting Vector can attest, that the equations can be hairy.  An equation like $\epsilon_0\frac{\partial}{\partial t} \left (E\times B \right)$ can quickly blow up into several terms, if you are not careful, especially if you don't recognize, at first, that certain collections of terms can be simplified into smaller quantities.  No matter how brilliant you are, if your algebra explodes enough, there will come a point where you will need some way of keeping track of all your expressions.  Not properly doing so risks the possibility of making an obscure algebraic error somewhere, which can be the source of much hair pulling.  This elicits the question of if there is a good way of notating your algebra so that you don't get lost.  Unfortunately, an internet search does not yield very useful results, especially since Google loves to substitute "complex" for "complicated," and we know where that leads!  Fortunately, I have worked out my own method that seems to work pretty well.  So, for this post, I will take a detour from discussing derivations to discussing derivation technique.  If anyone else has different methods to share, I would definitely encourage it in comments.

Issue 1: Introducing New Variables

Suppose that you have just worked out an expression, and it results in a long list of terms.  For example,

$(A + B + C + D)^2 = A^2 + B^2 + C^2 + D^2 + 2AB + 2AC + 2AD + 2BC + 2BD + 2CD$

Yes, I know that can be rewritten in concise form, using summation notation.  Shut up.

Now, suppose that you can expand each term further, but you are particularly interested in $2AB$, and wish to evaluate that first.

Well, one obvious solution is to next write,
$X = 2AB = ...$
And then proceed with the expansion.  Eventually, you substitute X back in.  That's not so bad, obviously.

What if, however, you want to eventually expand some terms, but combine others.  For example, you recognize that $A^2 + 2AC + 2CD$ means something.  Now, you have a problem, because your terms appear in different places in your sum.  In the example I gave above, it isn't so difficult if you write,

$Y = A^2 + 2AC + 2CD = ...$

and proceed as before.

Of course, this changes if your expression is much worse, say if you are raising to the 5th power instead of the 2nd.  Now, you have a problem, since your terms might be scattered all over the place.

You might solve this by reorganizing the terms in the equation to be close to each other, but that means writing your unfortunately long equation again.  There is also the issue that you are doing redundant work by writing, $A^2 + 2AC + 2CD$ again, just to remind yourself what Y is.

Additionally, there is another problem, which I have read others gripe about, and that is that we are about to run out of variable names.  We started at X.

The natural procession would then usually be Y, Z, then W, and then...what?  R, S, T, U, V?

You don't want to have variables that look like preceeding variables, and you don't want to just arbitrarily start assigning random letters, because the relationship between variable name and assignment would be...well...random.  Maybe you start using greek letters like $\eta$.  Now, you need to know the greek alpabet, and, on top of that, mathematicians and physicists love to use greek letters, so you will get collisions.

Fortunately, there is better shorthand to use.  For lack of a more imaginitive name, I call this "curly brace indexed assignment."  Now, before we get into this, keep in mind that this won't apply to certain situations in which there are collisions between notations and variable names, but I believe that in those cases, the modification of the notation is so straightforward as to not be worth mentioning.  It is also important to point out that this notation is a work in progress.  It is probably not complete and likely doesn't cover certain cases, even if you have solved a collision problem.  It is simply something that makes keeping track of your algebra in many cases easy.

Let's start with assigning X, since that's the simplest case.  First, X sucks because it is at the end of the alphabet.  So, let's start with A.  Of course, we see that A already exists as a variable, so let's find some way to easily distinguish it.  The method to do this should really depend on the situation.  In this case, I have simply chosen to index it with a subscript: $A_0$

That's where the "indexed" comes from in my name.  Now for the curly brace.

Under (or over) the same expression, I simply write this:
$(A + B + C + D)^2 = A^2 + B^2 + C^2 + D^2 + \underset{A_0}{\underbrace{2AB}}  + 2AC + 2AD + 2BC + 2BD + 2CD$

For the case of Y, obviously, it won't work to do this, because they expressions appear at different points in the sum.  But, that's ok.  You could simply do the following:
$(A + B + C + D)^2 = \underset{B_0}{\underbrace{A^2}} + B^2 + C^2 + D^2 + \underset{A_0}{\underbrace{2AB}}  + \underset{B_0}{\underbrace{2AC}} + 2AD + 2BC + 2BD + \underset{B_0}{\underbrace{2CD}}$

You might have expected me to increase the index, and choose $A_1$ for this assignment, but remember, you are trying to make this easy on yourself.  Indexes can quickly get out of hand as the number of your variables grows.  So, what you are actually doing is taking advantage of the fact that the alphabet has 26  letters.  That means that, assuming there are no collisions between existing indexed variables, and this notation, you get to assign 26 very different looking variables (or possible 52, if you also choose to use lower case), before you have to increase the index.

In this way, the index serves two purpose, as an endless source of variables, and a way to quickly distinguish new variables from preexisting ones.

Now, obviously, if you had an $X_0$ somewhere, you have a problem, but, this can easily be fixed, and again, it depends on the situation.  Perhaps you could place the subscript on the left side  ${}_0X$ instead, but that's just one solution.

Of course, the preceding examples assume that your sum can be partitioned into non-overlapping sets.  What if you recognize that $\frac{B^2}{2} + C^2$ and $\frac{B^2}{2} + D^2$  are interesting terms?  Here is one way to introduce variables:

$(A + B + C + D)^2 = \underset{B_0}{\underbrace{A^2}} + \underset{2C_0}{\underbrace{B^2}} + \underset{D_0 - C_0}{\underbrace{C^2}} + \underset{E_0 - C_0}{\underbrace{D^2}} + \underset{A_0}{\underbrace{2AB}}  + \underset{B_0}{\underbrace{2AC}} + 2AD + 2BC + 2BD + \underset{B_0}{\underbrace{2CD}}$

Then this implies
$D_0 = \frac{B^2}{2} + C^2$

Likewise,
$E_0 = \frac{B^2}{2} + D^2$

Alternatively, if you didn't want to deal with working out algebra in your head, you could write,
$(A + B + C + D)^2 = \underset{B_0}{\underbrace{A^2}} + \underset{C_0\left\{/2\right\}, D_0\left\{/2\right\}}{\underbrace{B^2}} + \underset{C_0}{\underbrace{C^2}} + \underset{D_0}{\underbrace{D^2}} + \underset{A_0}{\underbrace{2AB}}  + \underset{B_0}{\underbrace{2AC}} + 2AD + 2BC + 2BD + \underset{B_0}{\underbrace{2CD}}$

In this case, the commas in under the curly braces indicate two variable assignments, and the $\left\{/2\right\}$ signifies that you will be dividing by 2 before adding the term to the respective variable.
I haven't tried this yet, but for more complicated expressions, I would suggest using lambda calculus.  For example, if you wanted to take your summand and apply the polynomial, $x^2 - x + 4$ to it, you would write $\left\{\lambda x.(x^2 - x + 4)\right\}$

Now, this works for summation, but what about multiplication?  Consider the following example:
$E(A + B + C + D)$
Suppose, you are interested in A, but want to remember that it is multiplied by E.

Then, you would write
$\underset{A_0}{\underbrace{E}}(\underset{A_0}{\underbrace{A}} + B + C + D)$

Issue 2: Deferring stuff you don't care about.

Sometimes you are doing a horrifying expansion, but you are only interested in part of it.  You don't want to work out the rest of the expansion, at least, not yet.  You simply want to reason with part of it.

Then, you can use ellipses, in combination with curly brace indexed assignment.

Let's say that you have
$(A+B+C+D)^4$

This will have waaaay to many terms to write out if you are only interested in, say, ABCD.

The solution is to write the following:
$(A+B+C+D)^4 = \underset{A_0}{\underbrace{ABCD}} + \underset{B_0}{\underbrace{...}}$

In this case the ellipses stands for "everything else in this sum."

Any Other Ideas?

That concludes my PSA for the day.  Please let me know if you have any suggestions for improving this technique, or alternative techniques.

Wednesday, November 18, 2015

In Response to A Post on phys.org

So, this blog posting is not for everyone.  I wrote it in response to a comment made here by bschott. Initially, I was just going to post a comment, but it rapidly became too large to be a comment.  Thus, in the interest of not spamming phys.org, I will post it here, and link to it with a comment.  Here it is:

For some context, here is a posting by bschott elsewhere on the site:


Time dilation is a human perceptual artifact, not a physical reality. Assume the entire universe is your reference frame and you can watch anything happening in it at any time. The amount of "time" a series of events requires to occur, if they are the exact same events, will be the same no matter what your location or velocity is in the universe. Dilation only occurs between separate IRF's when comparing events in one to events in another. The effect disappears if you view everything simultaneously as opposed to each group in each RF measuring the ticking of a stopwatch and comparing it only to the other group.

Now, admittedly, there are far crazier things on the site, but it is certainly a demonstration of this poster's failure to understand fundamental concepts of Special Relativity.  In particular, saying "the entire universe is your reference frame" is meaningless.  A reference frame is a coordinate system.  The universe is not a coordinate system.

I mean, what does he mean?  Does he mean the center of gravity of the entire universe?  Well, we don't know that one exists, because we don't know how much universe there is.  Does he mean there is some preferred frame of reference in which everything "plays out"?  In that case, he's begging the question.  He would have to demonstrate that such a preferred frame exists in the first place, but we have no evidence of this.

Another problem is his assumption that even if you could find this frame of reference, causality implies that time ticks the same way everywhere in the universe, but this is just not correct.  First, in terms of SR, not every event is causally related.  Events outside the light cone are not causally related.  That means, depending on how you are moving such events can happen in different order. Furthermore, even for causally related events, time necessarily does not tick in the same way.  The whole point of General Relativity is it does away with the notion that there can be a universal rate of time in the presence of gravity.  Indeed, there are only very specific instances where this can be true: static space-time.  But I digress.  This guy doesn't know what he's talking about, and yet, saw fit to say the following:


Funny, I became a poster at this site due to the amount of crackpottery that is entrenched in mainstream theoretical physics and how militant it's supporters are here.

Concerning Non-Mainstream Theories

Now, there are three main kinds of non-mainstream physics, which are not mutually exclusive:

  1. a rigorously sound theory, at least based off of acceptable first principles, which can be falsified, or at least seems likely to be falsifiable in a few years
  2. a rigorously sound theory, which has been ruled out by conclusively by experiments
  3. a poorly defined theory that is not even wrong.  Either there is little or no evidence for it, it is based off of a faulty misunderstanding of establish physics principles, or it is simply too vague to be falsified.


An example of 1) would be Brans-Dicke theory
An example of 2) would be aether theory
An example of 3) would be intelligent design, and most personal hypotheses peddled by people on phys.org.

You will never find me vehemently denying a theory of type 1).  You might find me clarifying that it hasn't yet been demonstrated conclusively, but, if someone wants to provide a cogent, well research treatise on why they think Brans-Dicke theory explains a particular effect, go right ahead.

When it comes to 2 and 3, however, all bets are off.  At least 2, though, has the benefit of being formally defined, but we know better, and that means you need to get with the program.  It has nothing to do with militarism, and everything to do with not beating a stinking equestrian corpse splayed out on the ground with its tongue hanging out of its mouth.

...and if your theory is 3, well, what can I say?  I work on personal hypotheses all the time, and contrary to what posters of phys.org think, I certainly indulge in bucking the trend.  Indeed, I have even indulged in 2, when I have found pedagogical explanations unsatisfactory.  The difference is that, most of these hypotheses don't ever see the light of day because formalizing them usually reveals a fatal flaw in my thinking.  Even then, I would be very careful with revealing my hypothesis until I can find evidence for it that can't be attributed to something else, and this is something crackpots don't do.

Concerning Crackpot Posters


There are generally four types of posters I engage with on phys.org.

  1. is a poster who demonstrates general knowledge of science and peer review, and who tend to toe the line of mainstream science.  This poster typically provides well researched arguments whose pedigree can be easily traced to their original peer-reviewed source.  Posters like Thermodynamics, and Stumpy fall into these category.
  2. is a poster who demonstrates knowledge of science and peer review, but who does not toe the line of mainstream science.  In general, his/her arguments are also well researched.  I consider Noumenon to be an example of such a poster.
  3. is a poster who doesn't know much, but asks questions, and might submit personal hypotheses, but acknowledges that his/her hypotheses are possibly wrong.  This poster learns from mistakes, and sometimes can even offer insightful comments.
  4. is a poster who doesn't know much, and does not demonstrate competent understanding of any particular field, but regardless of that, insists that his personal hypothesis should magically trump all of science, evidence, and cogent arguments be damned.  They are resistant to learning, and react to any criticism with hostility.


These categories are not mutually exclusive either.

I have always been very respectful of any poster who falls into categories 1 to 3.  While they might not actually accept mainstream science, they all have one thing in common: they can be reasoned with.

OTOH, people who belong to 4 are impossible to reason with, inconsistent in their arguments, and never, EVER, willing to acknowledge that another argument has even been made, let alone how correct it is.  I liken arguing with them to catching fish with your bare hands.  You can never pin them down.  Their method of reasoning is so slippery and incoherent, it is impossible to get anywhere with them.  Show them evidence, and it is never evidence.  Provide them with a theoretical argument, and it's "just math."  Every argument you make with them always brings you back to square 1.  Ironically, these same people are also extremely predictable.  They invariably accuse mainstream science of rampant dogmatism, but are too incompetent to even recognize that the only reason for their own beliefs is dogmatism.

This brings us to people like bschott and Benni.  You are of type 4.  Hence, you believe that people who correct you only do so because they are militant, just as you are.  It doesn't matter that you can't tell me what Ricci curvature is, or what it means for gravity to be locally equivalent to acceleration.  It doesn't matter how many experiments confirm gravitational time dilation.  It doesn't matter, because you already decided, a-priori, that your vague understanding of SR and GR is incompatible with your world view.

You are the dangerous ones, because you act like authorities on subjects you don't understand, and your pseudoscientific prattling is just more evidence to sane people who fall into category 3) that scientists have their heads in the clouds.  So, if we seem militant to you, remember, it isn't because we are dogmatic.  It is because you are.

More Garbage

In response to my quote,
if it were up to me, I'd get rid of the entire voting system, but then again, if it were up to me, I would implement a system closer to Stack Exchange's, where you aren't allowed to post freely until you've demonstrated some kind of competence.
bschott said

LMAO...holy shit on a stick! Someone who is trained in mainstream physics is therefore competent...and is also(this is the best part) part of a group of people who math has convinced that 4/5 of the matter in the universe exists outside the EM spectrum. 

Now, I hate the term straw-man, mostly because it gets bandied about by people who don't understand what the term means (hint: it doesn't mean paraphrasing what you actually said or believe), but this is a bonafide straw-man.  We are not convinced by math.  We are convinced by demonstrated theory and evidence.  That means we model our conclusions on two things:

  1. Is there empirical evidence that a claim is true or false?
  2. In lieu of evidence, is there theoretical justification that a claim is true or false grounded in firmly established principles?
So, math has not convinced us.  Evidence, and theoretical arguments grounded in evidence have convinced us.

One thing I love, in particular, is when crackpots make assumptions about my dogmatism.  For example, they assume that I hold String Theory to same level of esteem that I do, say General Relativity.  Of course, nothing could be further from the truth.  String theory, at this point, is not even wrong in a variety of ways.  To me, it does not qualify for 1 or 2 because it is not grounded in firmly established principles, and certainly is not supported by empirical evidence.

What crackpots like bschott don't understand is that unlike String Theory, General Relativity is grounded in firmly established principles.  It's just that it is all too often popularly presented as if Einstein simply cut the notion that gravity is indistinguishable from curvature in space-time from whole cloth.  This is simply not true.  On the contrary, Einstein's insight was developed over several years, and it wasn't really until he started thinking about how we would measure the value of $\pi$ in rotating reference frames, that he came to this conclusion.  He didn't start with this hypothesis.  He reached it by doing lots and lots of work, and finding a pattern.  And even then, he still needed the expertise of people like Levi-Civita to help him formalize his ideas.

bschott also asked me,


Define competent please, because in the real world it means you didn't fuck up this badly.
It depends on the context.  In the case of Stack Exchange, competence is determined by the value of the questions you initially post.  If your questions don't have value, then you aren't likely to give valuable answers.

In the case of phys.org, ideally, I would want at least general familiarity with the concepts of peer review and the evidence available for hot-button topics.  Understandably, this would be difficult to enforce, so I would at least require an example of the candidate's prose in response to a popular hot-button issue to determine if they really understand what an acceptable comment should be.

A comment should really be about the article itself.  It should be specific, and well researched.  In writing your initial salvo, you should ask if it actually pertains to the content of the article, rather than a general ideological stance.

Therefore, if the article is about the effects of global warming on a certain species of Coral, your comment should actually refer to the effect, and not to whether global warming is a hoax and how much Al Gore is being paid.

If you cannot even write a well-researched response to such an issue, there's no reason to expect you to offer comments of any value in a forum that is historically unmoderated.

I might also require a candidate to provide a sample response to a deliberately inflammatory comment, and check to see if the response is measured and offers anything but more flaming.  Ironically, I might actually not allow myself to comment because I have occasionally offered only comments with the intent to wound rather than to contribute meaningfully to the conversation.  In fairness, however, sometimes this has been in response to people who can't otherwise take a hint that they are being pricks.


Well, that's all for now.

Tuesday, October 13, 2015

A Quick Start Guide to the Geometric Product - A Beginner's Perspective Part I

I am just learning about a phenomenal way of manipulating vectors and transforms called Clifford Algebra (or Geometric Algebra).  To give you an idea of how cool this, I first encountered this in reading a paper by David Hestenes on the Dirac Equation found here: (http://aflb.ensmp.fr/AFLB-283/aflb283p367.pdf), a paper I found by searching for the phrase "Dirac equation insight".  You see, the Dirac Equation is one of the closest things to black magic you will ever find in all of physics literature.  Now, as I am aware that some of the imaginary people who read this blog probably aren't well versed in Quantum Mechanics, I won't get into it, but I'll just say that he took an equation that uses bewildering abstract concepts, (which, in turn came from a formalism, that, though, esoteric, starts making sense once you strike your head forcefully, and in a disciplined manner, against hard surfaces enough) and reformulated it in a more concrete geometrical way.  In fact, one thing he was able to derive from his new reformulation was an interesting kinematic interpretation of electron Spin.  Now, don't be fooled by it being it being concrete and geometrical.  Indeed, to work his magic, Hestenes decided to rederive the equation using Geometric Algebra, which I will briefly describe in a little while.  As a caveat, I am a (possibly perpetual) theoretical math n00b (as you probably already realized), so if you are looking for a rigorous formulation of Geometric Algebra replete with axioms, and whatever the hell things like
 Gr_F C\ell(V,Q) = \bigoplus_k F^k/F^{k-1}
are, you'll have to look elsewhere.

Cramer's Rule: An Example to Whet Your Appetite

Now, unlike mathematicians, who almost universally seem to subscribe to the idea that gavaging students with definitions and theorems as an acceptable form of pedagogy, I will introduce you to example of how one might start thinking in a way that naturally leads to geometric calculus.  This section is actually quite long (as I need to account for deficits in understanding from my imaginary readership) so if you want to get to the actual quick start guide, skip it.

Let's start with a 2D system of equations.

$a_{00}x_0 + a_{01}x_1 = b_0\\
a_{10}x_0+a_{11}x_1=b_1$

Those of you familiar with linear algebra will think "no sweat", and and do something like treat the $a_{ij}$s as a matrix, M, treat this as a matrix equation, and then multiply both sides by the matrix inverse $M^{-1}$, which, of course, requires knowing what its components are.  Note, however, the similarity with the 1D case.  Ultimately, you end up multiplying both sides by an inverse.  This is a very consistent way to think about things, and with all linear equations, we can always do this, as long as we actually know how to find the inverse, but therein lies the rub.  It can be calculated, but doing so is laborious, and the end result makes little sense if you don't understand geometrically what's going on.

Now, there are other ways of solving  this system that are more geometrical.  I am going to use cross products.

First, we'll express this equation as 3D vectors:
$\begin{pmatrix}
a_{00}
\\
a_{10}
\\
0
\end{pmatrix}x_0 +
\begin{pmatrix}
a_{01}
\\
a_{11}
\\
0
\end{pmatrix}x_1 =
\begin{pmatrix}
b_0
\\
b_1
\\
0
\end{pmatrix}
$

Then, we'll just dispense with components, by writing

$\mathbf{A_0}x_0 + \mathbf{A_1}x_1 = \mathbf{B}$

Now, cross both sides by $A_1$ to solve for $x_0$

$\mathbf{A_1}\times\mathbf{A_0}x_0 = \mathbf{A_1}\times\mathbf{B}$

Then, we recognize that both sides point along the z axis, and so, we can dot with $\hat{z}$, and then divide by $x_0$'s coefficient on the LHS to get

$x_0 = \frac{(\mathbf{A_1}\times\mathbf{B} )\cdot \mathbf{\hat{z}}}{(\mathbf{A_1}\times\mathbf{A_0} )\cdot \mathbf{\hat{z}}}$

Likewise,

$x_0 = \frac{(\mathbf{A_0}\times\mathbf{B} )\cdot \mathbf{\hat{z}}}{(\mathbf{A_0}\times\mathbf{A_1} )\cdot \mathbf{\hat{z}}}$

Writing them out explicitly,

$x_0 = \frac{-a_{01}b_0 + a_{11}b_1}{-a_{01}a_{10} + a_{11}a_{00}}$
$x_1 = \frac{-a_{10}b_0 + a_{00}b_1}{-a_{10}a_{01} + a_{00}a_{11}}$

Now, if you know linear algebra, you will see that the denominators of $x_0$ and $x_1$ are just the determinant of the matrix

$\begin{pmatrix}
a_{00} & a_{01} \\
a_{10} & a_{11}
\end{pmatrix}$

Moreover, if you replace the first column of this matrix with the "b" vector,
$\begin{pmatrix}
b_0
\\
b_1
\end{pmatrix}$

You get the numerator of $x_0$.

Likewise, if you replace the second column of this matrix with the "b" vector, you get $x_1$'s numerator.

This is called Cramer's Rule, and is one of those things you can't believe nobody told you before.  It actually applies to linear systems of any dimension, and makes I (and many others) prefer it to the more laborious, and less illuminating, process of Gaussian Elimination, which is basically just the high school method of solving linear equations in slightly more respectable mathematical clothing.  Furthermore, Cramer's Rule has a geometric interpretation we can see, here.  Remember that the determinant is just a signed volume.  Therefore, Cramer's rule tells us that the solution to a linearly independent N degree linear equation is just the ratio of two different signed volumes.

The Geometric Product Between 2 Vectors


It turns out that the methods we have been using are precisely the methods that inspired the development of Geometric Algebra.  Now, dot and exterior products are both forms of multiplication, in that they distribute over addition, but they have some undesirable algebraic properties.  For example, the dot product takes two things (vectors), and returns an entirely different thing (a scalar).  In other words, it isn't closed.  Cross Product, while closed, is neither associative, nor commutative.  Neither operation has a useful concept of inverse.

William Kingdon Clifford had the brilliant idea to take the cross product and the dot product and combine them into one product that, in a sense, behaves better than either of them.  What I mean by this is that theoretically, it is more well behaved than either dot or cross product.  It is closed, unlike the dot product, and it is associative.  Furthermore, it allows for many elements to have inverses, and those inverses make geometric sense.  The price for this unification, is that the elements are no longer the familiar vectors and scalars of linear algebra, but are a collection of elements known as "multivectors", or "cliffs", depending on the author, but have no fear.  These elements can be represented as linear combinations of geometrically meaningful elements called blades.

Unfortunately, the Geometric Product is not that simple to present all at once, at least not if you don't want to assault the student with a barrage of new concepts, so we will start small.  So, without further ado, here is the geometric product between two vectors:

$UV = U\cdot V + U\wedge V$

It contains the $\wedge$ operator, which we have not defined, so let's do so.

As I said, the idea is to combine the dot product, and the cross product into a single multiplication.  However, there is are problems with doing that, the largest of which is that the cross-product is not associative.  In order for UV to be associative, $U\wedge V$ whatever it is, also must be associative.

To account for this, we introduce an entirely new operation that is both associative, and encodes the cross-product (or exterior product, for higher dimensions).

$U\wedge V$ is known as the wedge product, and its result is neither a vector, nor a scalar, but something entirely different.  There are two ways to think about it.  First, you can think about it as the parallelogram formed by U and V, with what's called an orientation.  An orientation can be clockwise or counter-clockwise.  Intuitively, imagine drawing out the parallelogram without lifting your pencil from the paper, using U as the starting edge, and V, as the edge.  This forces you to draw clockwise, or counter-clockwise.  If you draw counter-clockwise, your orientation is positive, and if you draw clockwise, your orientation is negative.  This is actually just the right hand rule.  It also means that the wedge product between two vectors encodes the cross product.

There is another way of speaking about this orientation, alegbraically, and that is that the wedge product between two vectors is anti-commutative.

Thus,

$U\wedge V = -(V \wedge U)$

The second way to think about the wedge product is as the subspace spanned by U and V.

The wedge product between two vectors is known as a 2-blade.

Now, some of you might be puzzled by the presence of this wedge, which is not a real number, and the dot, which is a real number, so we are adding un-like objects.  In fact, in geometric algebra, a general object can contain vectors, too!  However, we've added un-like objects to each other for years with complex numbers, so as long as this sum is irreducible to a real or bivector, there is no reason to complain.

So, what happens when we wedge $U\wedge V$ with a third vector, W?  Unsurprisingly, you can think of

$U\wedge V \wedge W$

as the parellelpiped formed by U, V, and W, again with an orientation, which is either -1, or 1, depending on orientations of the faces of the parallelogram.  The way to envision this orientation geometrically, this time, by drawing the outline of the parallelpiped, by drawing U as the first edge, V as the second edge, and W as the third edge, as shown in Fig. 1 (set U = P, V = Q, and W = R).  Unsurprisingly, this is called a 3-blade.  Likewise, wedging 4 times creates a 4-blade, and wedging k times, creates a k-blade.  Not everyone uses this terminology.  It is very common, also, to refer to a k-blade as a multivector, with a 2-blade known as a bivector, and a 3-blade known as a trivector.  I prefer the former terminology because it is less ambiguous, as using "vector" might suggest that the elements in question are types of vectors, which, they aren't at least not in a pedagogical sense.  Again, the other way to think about a k-blade is as the subspace formed by the k-wedged vectors.  The dimension of a k-blade is known as its grade.

Figure 1.  Source: https://slehar.wordpress.com/2014/03/18/clifford-algebra-a-visual-introduction/
Now, in introducing the concept of a k-blade, I lied a little.  It isn't always true that wedging with a vector with a k-blade produces a (k+1)-blade.  Geometrically, this is easy to see.  If the vector is a member of the subspace spanned by k former vectors, then you won't increase the dimension.  Another way to think about this is that the new parellelpiped will have a k-dimensional volume of zero--in other words, it vanishes!  Furthermore, we want to relate this to the cross-product, which we definitely want to associate with the 2-blade.  Therefore, if $U_0$, $U_1$, ..., $U_{k-1}$ are a set of linearly dependent vectors,

$U_0 \wedge U_1 \wedge \text{...} \wedge U_{k-1} = 0$

Let's return to the geometric product.

Now, one cool thing about this definition is that we can then recast dot and wedge in terms of this new multiplication.  Namely, dot product is

$\mathbf{U}\cdot\mathbf{V} = \frac{\mathbf{U}\mathbf{V} + \mathbf{V}\mathbf{U}}{2}$

and wedge is

$\mathbf{U}\wedge\mathbf{V} = \frac{\mathbf{U}\mathbf{V} - \mathbf{V}\mathbf{U}}{2}$

The geometric product encodes a lot of operations we perform with matrices.  As such, it simplifies thinking about things in several dimensions.  If you know it, it makes things like deriving Cramer's Rule trival.

I will give you a taste of this, now.  Take each side of your 2D system of equations and wedge it with $A_1$, to get,
$(\mathbf{A_1} \wedge \mathbf{A_0})x_0 = \mathbf{A_1} \wedge \mathbf{B}$

It turns out that  $(\mathbf{A_1} \wedge \mathbf{A_0})$ has a multiplicative inverse, which we will see in my next posting, so, we get

$x_0 = \frac{\mathbf{A_1} \wedge \mathbf{B}}{\mathbf{A_1} \wedge \mathbf{A_0}}$

Once you know how to find the multiplicative inverse, then all you need to do is plug it in, and you will arrive at your answer.

What's the *%@!!#!!! Definition of Geometric Product Already?

[Update: I think I now have a short definition of the Geometric Product, though, it requires knowing about, and understanding the extension of the inner product to multiple vectors.  Watch for the next posting for this definition.  I must say that after working with Geometric Algebra more intensely over the last week or so, I am strongly suspecting that, contrary to what I have been reading, the geometric product is a fairly simple concept to understand, and does not require lists and lists of definitions and theorems just to establish the key concepts.  On the contrary, I am almost certain that, as usual, the requisite key concepts are few, and easy to understand.]  


Unfortunately, there are several short answers, each with its own advantages and disadvantages.


  1. The geometric product is the associative multiplication you must get when you have a vector space, and your only further assumption (plus some minor technical assumptions for consistency's sake) is that a vector times itself is its squared magnitude.
  2. The geometric product is a product between linear combinations of blades and returns another linear combination of blades.  If you multiply two blades, the the resulting linear combination can be organized into geometrically meaningful groups, all sharing the same grade.
  3. The geometric product is the multiplicative operation of the ring generated by scalars and vectors, where a vector times itself is its magnitude squared.
  4. The geometric product is given by the following formula:
Source: http://arxiv.org/pdf/1205.5935v1.pdf
These are probably not very enlightening.  In fact, definition 1 was downright shocking to me.  I actually spent a few days convinced that it was an incomplete definition, and that you needed one other axiom concerning the definition of wedge products to have a complete, theory--but nope! [Update: I had said that you don't need an axiom concerning the definition of wedge products, and depending on your pedagogical needs, this isn't necessarily correct.  Personally, I do think that this axiom is required.  I made my initial claim based on a mistake I made in deriving the general wedge product ] Definition 2 is not really a fundamental definition, but it is at least descriptive, and prepares us for how to start thinking about the geometric product.  Definition 3 is the closest that I will come to giving a formal definition of the geometric product.  As for definition 4...no...just...no.

...

Ok, fine, but don't say I didn't warn you.  Definition 4 is the geometric product between two blades in all all its glory. You can understand it this way: Definition 4 gives a precise equation describing Definition 2.  It introduces the unfamiliar angle-bracket notation, where

$<A_rB_s>_k$ refers to generalizations of the wedge and dot product with grade k.  Specifically,

$<A_rB_s>_0 = A_r \cdot B_s$, and yes you can take the dot product between general blades.

and

$<A_rB_s>_2 = A_r \wedge B_s$

However, at this point, it would be silly to say more about them, and I only included it to show that there is a formula for the geometric product in terms of geometrically meaningful operations.

Unfortunately, none of these definitions are clear enough for my tastes.  It wish I had something as precise and illuminating as "multiplication between two numbers, M and N, is M added to itself N times", but alas, at this point of my understanding, I cannot.  At some point, in the future, when I have a fuller understanding of this subject, I hope that I can, despite Alan Macdonald's frustrating insistence to the contrary.  However, that doesn't mean I can't give a straightforward, albeit, long, description of it by building up concepts.

In fact, whether you realize it, or not, we have already defined the geometric product.  No other geometric product is possible, given the rules we have already established.  It naturally falls out.  The reason for this is our requirement of associativity of the product.  However, I won't leave you high and dry, or this wouldn't be called a "quick start guide."

The Geometric Product for Vectors

Now, I remarked that the geometric product relates two blades to each other in geometrically meaningful ways.  Superficially, you should already have seen this with my definition of the geometric product between two vectors being the sum of the dot and wedge product between them.  But to show you that I am not out of my gourd, and there really is computational power to this product, we'll explore it in more detail.  Let's consider a really concrete example: vectors in 3D space.

The unit vectors are $e_0$, $e_1$, and $e_2$.

Then, the geometric product for two vectors is pretty easy to see.
Namely,
$e_ie_i = {e_i}^2 = 1$

and

$e_0e_1 = -e_1e_0 = e_0\wedge e_1 = -e_1\wedge e_0$
$e_1e_2 = -e_2e_1 = e_1\wedge e_2 = -e_2\wedge e_1$
$e_2e_0 = -e_0e_2 = e_2\wedge e_0 = -e_0\wedge e_2$

In fact, if any two vectors, U, and V are orthogonal, then $UV = U\wedge V$.  This is easy to see, since we have already defined what the geometric product is for two vectors.

So, immediately, from this, we can see that the geometric product of two vectors
$a_0\mathbf{e_0} + a_1\mathbf{e_1} + a_2\mathbf{e_2}$

and

$b_0\mathbf{e_0} + b_1\mathbf{e_1} + b_2\mathbf{e_2}$

is
$a_0b_0 + a_1b_1 + a_2b_2 + (a_0b_1-a_1b_0)\mathbf{e_0\wedge e_1} +  (a_2b_0-a_0b_2)\mathbf{e_2\wedge e_0} + (a_1b_2-a_2b_1)\mathbf{e_1\wedge e_2}$

One important thing to realize about this expression is that the wedges are interchangeable with the geometric product, as long as we are wedging orthogonal vectors.  So, we can replace this expression with,

$a_0b_0 + a_1b_1 + a_2b_2 + (a_0b_1-a_1b_0)\mathbf{e_0e_1} +  (a_2b_0-a_0b_2)\mathbf{e_2e_0} + (a_1b_2-a_2b_1)\mathbf{e_1e_2}$

The coefficients of the $e_ie_j, i \neq j$ blades should look familiar.  They are precisely the components of the cross-product!  In other words, the cross product is actually encoded in the geometric product between two vectors.  This is no accident; it is a result of how we have defined the wedge product. [Update: Holy ^!%#%! and I just figured out how to recover the cross product (and the generalization of this, the exterior product) from this--and it is exceedingly simple.  I will cover this in  a blog posting, or two]

Breaking vectors into orthogonal components sheds light on what the product is, and is the first strategy I used in attempting to understand this algebra.  However, there is another way to understand it, using permutations, that is remarkably more effective for establishing all of its identities.  However, let's just stick to orthogonal vectors for now.  Sets of orthogonal vectors hold a privileged position in geometric algebra, as proven in our first theorem:

Anticommutivity Theorem:

Suppose we have n distinct, orthogonal, vectors, and you multiply them: $v_0v_1v_2...v_{n-1}$.
Well, it turns out that this product must be anti-commutative.

Proof:

To see this, we use associativity, which shouldn't be controversial, since we are assuming our product is a form of multiplication.

Now, without loss of generality, pick $v_0$ and $v_1$ and swap them.
This yields
$(v_1v_0)v_2...v_{n-1}$


Furthermore, we know that the geometric product of two orthogonal vectors is interchangeable with the wedge product, so, we get
$(v_1\wedge v_0)v_2...v_{n-1} = -(v_0\wedge v_1)v_2...v_{n-1} = -v_0v_1v_2...v_{n-1}$

In fact, the geometric product of orthogonal vectors takes a particularly simple form--a k-blade.  Now, initially, when I wrote this posting, I had thought to make this an axiom, and then, I made a mistake, and reneged on it.  However, I see, now, that my initial suspicion seems to be correct--in a sense.  You see, there are two ways of defining the wedge product in association with the geometric product.  The first is to define it explicitly in terms of the geometric product, which, despite being an awful pedagogical approach, is favored by pretty much everyone.  The second is the way I use, which is to define it, and then base the geometric product off of it.    Admittedly, the first approach is more elegant, in that you need fewer axioms, but I don't know why mathematicians (and many physicists) think that elegance is a good substitute for understanding.  It doesn't matter a whit how concise and direct your argument is if you can't get your point across to all but a subset of people who enjoy the gavage method of learning concepts${}^{\text{TM}}$.  Seriously, dudes, I know I harp on this a lot, but guys, how about you tone down the rigor when presenting new concepts?  You might get more people to like your subject, if you do.

Anyway, I keep meaning to make a blog post about my love/hate relationship with theoreticians, but that's for another day.  For now, let's just add an axiom, which I will call the

Fundamental Axiom

If $v_0$, $v_1$, ..., $v_{k-1}$ are all orthogonal vectors, then,

$v_0v_1...v_{k-1} = v_0 \wedge v1 \wedge ... \wedge v_{k-1} $

The motivation for this is consistency (as far as I can tell.  I am not certain what the original motivation was).  For two orthogonal vectors, this is true, so why not have it be true for more than two?  Indeed, there is a more formal reason why, based off of redundancy, but that is too complicated to get into, now.

Anyway, once we have this axiom, the geometric product can be completely defined.  By that, I mean you could write a straightforward, albeit, inefficient algorithm for calculating the geometric product using wedge products and scalars.

To see this, first, I need to show that if you have a product of multiple vectors, its value will depend fundamentally on how we define the product of orthogonal vectors.
Let's say you have a set of orthogonal vectors, O of size m, and you have a product that contains N vectors from this set O, where n > m.  Then, by the pigeonhole principle, this product contains at least one copy.

Let's construct two sets: $C_{odd}$ and $C_{even}$.

$C_{odd}$ is the set of all vectors that occur an odd number of times in the product.
$C_{even}$, likewise, is the set of all vectors that occur an even number of times in the product.

Furthermore, let $k = |C_{odd}|$

Orthogonal Product Theorem

Let $v_0$, $v_1$, ..., $v_{k - 1}$ be the set of vectors in $C_{odd}$.  Then, this product will be equivalent to $s\mathbf{v_0v_1}...\mathbf{v_{k-1}}$, where s is a scalar.

Proof:

I will start with an example.  Let O be {A,B,C,D}.  Suppose you have the product, $AAABBBCD$.
Rearrange this so you have $ABCD$ on the far right:

$AAABBBCD = AABBABCD$.

Now, $AA = A^2 = a$ and $BB = B^2 = b$, where a and b are scalars.

So, the result is:
$abABCD$

So, for the general proof, just as we did for the  we are going to reorganize the product in the following manner, multiplying it by the appropriate sign:
Let the right-most k terms be $v_0v_1...v_{k-1}$.
In the remaining product, there will be vectors in $C_{even}$, and vectors in $C_{odd}$.  Take each set of copies and place them to the left of $v_0v_1...v_{k-1}$ so that all copies of the same vector are adjacent to each other.  Now, the adjacent "runs" of $C_{even}$ will each yield scalars.  Likewise, the remaining adjacent "runs" in $C_{odd}$ will also produce scalars, since they will be even in length.  Then s is the sign multiplied by all the scalars to the left of $v_0v_1...v_{k-1}$.  We have thus proven our theorem.

Therefore, if we are only using orthogonal vectors, the result will simply be a scaled version of whatever the geometric product is for a set of distinct orthogonal vectors.''

Before I go on, let me introduce an abuse of terminology to make my life easier.  Call, a scalar, a single vector, or a product of multiple, distinct, orthogonal vectors a blade.  If it has k elements, then, we can also call it a k-blade. Furthermore, call a product of orthogonal vectors, which might include copies an orthogonal product.

Now, let's use Theorem 2 to prove its general version:

Geometric Product Theorem

Every geometric product of vectors is a linear combination of distinct blades.

Proof:

The proof for this is straightforward, but tedious if you write it out with actual algebra.  For the sake of brevity, I will merely give a prescription for how to do this.  Every vector is a linear combination of basis vectors.  Now, distribute the product over all linear combinations.  Then, you will get a sum of orthogonal products.  By theorem 2, each of these orthogonal products are then simply a blade multiplied by some scalar.

Thursday, March 19, 2015

Proof that Objects Don't Always Return Under Gravity

So, I told this crackpot over at phys.org that I would be proving what is written in the title.  It is now late, and the proof is turning out to be really long, really because he's a bit of a weasel, and thinks he can get away with saying vague, hand-wavy, things like, "Heisenberg kicks in when velocity gets low enough."  Here is the proof, so far, so that Mr. Crackpot doesn't think I am sitting on my hands.  NOTE: THIS PROOF IS NOT FINISHED, YET: [UPDATE:  Here's some background for this blog posting.  Over as phys.org, I have a long history of arguing with a crackpot I call Clever Hans.  That isn't his screen name, but a nickname that I gave him after realizing that his approach to understanding math and physics was about as deep and methodical as the approach taken by his namesake.  This is not the first time I have attempted to prove this particularly well known property of gravity to him.  My first attempt, however, involved solving differential equations and using limits, something he summarily dismissed as "only a mathematical model," adding that at some point, "Heisenberg would kick in," thereby demonstrating not just a poor understanding of quantum mechanics, but an even poorer understanding of potentials.  Anyway, to make a long story short, I decided to attempt this task once again, this time with just high school algebra, conservation of energy, and conservation of momentum.  I don't know why I bother, though.  I am sure that his response will be something profound, like, "you forgot a comma, therefore, your proof is invalid."  ]

All right, Mr. Crackpot.  Thank you for your patience. Here is my proof.

Before I start, let me establish something about gravitational potential.
Now, the gravitational potential is given by -GmM/r, according to http://hyperphysics.phy-astr.gsu.edu/hbase/gpot.html, where r is the distance between the gravitating objects, and m and M are the masses.

But, you have insisted, often, that the laws of gravity as presented by textbooks, are merely approximations.  Therefore, I will allow for the more general potential, $\frac{-GmM}{r} + s\left(r\right)$, where s(r) is an error correcting term.  Now, at normal distances, s(r) must be approximately 0 because, otherwise, we would have detected it by now.  However, we are concerned with what will happen at large distances, and so, s(r) may come to dominate the potential.  More on that in a little bit.

Now, you argue that, in a universe with only two gravitating objects, they will eventually return to each other, regardless of their initial velocities.  Normally, I would say this means that at least one of the objects will slow to 0m/s, and then change direction.  However, I anticipate that you would counter that some kind of hypothetical quantum effects might allow the object to "jump" between moving forward and moving backward without completely slowing to a halt, and so, I'll grant you that, for argument's sake.  Let's say that whenever an object slows, and turns around relative to another object, there must be some lowest speed threshold, $S_q$, such that AT SOME POINT BEFORE TURNING AROUND, the relative speed will be less than or equal to $S_q$.

For example, let's say $S_q = 10^{-12}m/s$.  Then the relative speed between the two objects must at least slow down to $10^{-12}m/s$, if not less, before at least one of the objects change direction.  However, as with s(r), let's not assign it a value just yet.

With this in mind, all I need to do to prove you wrong is find a counter example to your claim.  So, here is what I will do. I will do a proof by contradiction, and assume that you are correct, and then show that it implies physically unreasonable conditions for a specific case.

For this, I have chosen a very simple set up.  We choose two identical, spherical, objects, A and B, of mass, 10^13kg, about the mass of the comet, 67P/Churyumov-Gerasimenko.  Object A starts out at $-10^{6} $ m from the origin.  B starts out at $10^{6}$ m from the origin.  Their initial velocities are equal in magnitude and in opposite directions, so that A is travelling, initially at $-v_0$, and B is travelling initially at $v_0$.  I'll get to the value of $v_0$ in a little bit.  For now, let's at least require that $|v_0|$ is well below the speed of light so that relativistic effects are negligible.  With this in mind, let's calculate the gravitational potential of A and B for this initial condition.

Well,

\begin{equation}-\frac{Gm^2}{2\times 10^6 \text{meters}} = -3369200000 \text{J}\end{equation},

 so, our potential is $-3369200000 \text{J} + s(2\times 10^6 \text{meters})$.

Now, let's see what conservation of momentum tells us.  Well, the initial total momentum is $mv_0 - mv_0 = 0 \textrm{kg m/s}$.

That means that if vA and vB are the velocities of A and B respectively in the future, m*vA + m*vB = 0, so that vA = -vB.  So, again, vA, and vB will be equal in magnitude an opposite in direction.  Therefore, we can dispense with this notation, and instead assume that at some point in the future, A is travelling at $-v_1$, and B is traveling at $v_1$.  Currently, we don't know what $v_1$ is.  It could be negative or positive or 0.

Suppose that A is at position xA and B is at position xB.  Since A and B are always moving at the same speed, we can also assume that xA = -xB.  As above, I can dispense with the "A" and "B" notation, and instead say that A is at position $-x_1$, and B is at position $x_1$.  Again, we DON'T YET KNOW what $x_1$ is.

Without loss of generality, assume that B turns around at some point in the future.  Now, since the initial conditions are that A and B start out moving away from each other at the same speed, we can assume that A and B will have moved even further away from each other by the time that their relative speed, $2x_1$, slows down to Sq or lower.  So, we can DEFINITELY assume that this happens when B's position, $x1$ is $> 10^6 \text{meters}$.

Now, let's use conservation of energy to get an idea of how large $x_1$ needs to be so that its magnitude is between 0meters/s and $S_q/2$.

The initial total energy is

$mv_0^2 - 3369200000 \text{J} + s(2\times10^6 \text{meters}) = (10^{13 }\text{kg})\times v_0^2 - 3369200000 \text{J} + s(2\times 10^6 \text{meters})$

Well, $\frac{-Gm^2}{2x1} = -\frac{3.3692\times10^{15} \textrm{kg meters}^3/\text{s}^2}{x_1}$
So, the total energy is $(10^{13}kg)\times v_1^2 -(3.3692\times 10^{15} \textrm{kg meters}^3/ \text{s}^2)/x1 + s(2x1)$
and conservation of energy tells us that
$S_q^2 = v_1^2 = v_0^2 - 0.00033692 \text{J\kg} + \frac{s(2\times 10^6 \text{meters}) }{10^{13}\text{kg}}+ \frac{ 3.3692\times10^{15} \textrm{kg meters}^3/\text{s}^2}{x_1\times10^{13}\text{kg}} - \frac{s(2x_1)}{10^{13} kg} = v_0^2 - 0.00033692 \text{J/kg} + \frac{s(2 \times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{336.92 \text{meters}^3/\text{s}^2}{x1} - \frac{s(2x_1)}{10^{13}\text{kg}}$

Or more succinctly,
$S_q^2 = v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{ 336.92 \text{meters}^3/s^2}{x_1} - \frac{s(2x_1)}{10^{13}\text{kg}}$

Now, there are two cases to consider: that s(r) is bounded as r increases, or that s(r) is unbounded as r increases.  I assert that it is physically unrealistic if s(r) is unbounded as r increases.  I'll assume that Mr. Crackpot agrees with this since his argument is that as distances grow, quantum effects "kick in", meaning that s(r) becomes quantum energy fluctuations.  In other words, the only two reasonable contributors, in this non-relativistic case, to the gravitational potential energy are the classical gravitational potential energy and quantum energy fluctuations.  Now, I am aware that Mr. Crackpot, being unabashedly bad at physical intuition, might, indeed, protest that I haven't considered the case where s(r) is unbounded.  If he does, I have a proof for that, too, but in the interest of not making this post any more of the novella it is, I will assume that it is acceptable to require that s(r) is maximally bounded.

Thus, at most, s(r) is bounded in magnitude by a particular energy scale.  Let's call this scale $s_{ub}$, so that $-s_{ub} \leq s(r) \leq s_{ub}$.  Note that $s_{ub}$ is nonnegative.

With this in mind, we can say that

$\frac{s(2x_1)}{10^{13}\text{kg}} = v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{ 336.92 \text{meters}^3/s^2}{x_1} - S_q^2$

implies the inequalities,

$-\frac{s_{ub}}{10^{13}\text{kg}} \leq v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{ 336.92 \text{meters}^3/s^2}{x_1} - S_q^2 \leq \frac{s_{ub}}{10^{13}\text{kg}}$

We only need the right inequality,

$v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{ 336.92 \text{meters}^3/s^2}{x_1} - S_q^2 \leq \frac{s_{ub}}{10^{13}\text{kg}}$

This is equivalent to writing

$\frac{s_{ub}}{10^{13}\text{kg}} +  S_q^2 \geq v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}} + \frac{ 336.92 \text{meters}^3/s^2}{x_1}$

However, since $x_1$ is positive, it is also true, if we just simply drop the term containing it.

$\frac{s_{ub}}{10^{13}\text{kg}} +  S_q^2 \geq v_0^2 - 3.3692\times 10^{-4} \text{J/kg} + \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}}$

Now, I will reveal the value of $v_0$.  I have assigned it the hugely nonrelativistic value of 340.29 meters / s.

So, we have

$\frac{s_{ub}}{10^{13}\text{kg}} +  S_q^2 > 115797.2841 \text{meters}^2/\text{s}^2 - 3.3692\times 10^{-4} \text{J/kg}+ \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}}$

which is equivalent to

$\frac{s_{ub}}{10^{13}\text{kg}} +  S_q^2 > 115797.28376308\text{meters}^2/\text{s}^2+ \frac{s(2\times 10^6 \text{meters})}{10^{13}\text{kg}}$


Now, $2\times10^6\text{meters}$ is not very large.  It's in the neighborhood of 1000km. We know, then that $s(2\times 10^6 \text{meters})$ is too small to detect, here at home, so that we can simply say that its value, compared with $115797.28376308 \text{meters}^2/\text{s}^2$, is approximately 0.  It is, therefore, acceptable to write,

$\frac{s_{ub}}{10^{13}\text{kg}} +  S_q^2 > 115797.28376308 \text{meters}^2/\text{s}^2$

Now, the above inequality implies that

\begin{equation}\frac{s_{ub}}{10^{13}\text{kg}} \geq 57898.64188154\text{meters}^2/\text{s}^2\end{equation}

OR

\begin{equation}S_q^2 \geq 57898.64188154 \text{meters}^2/\text{s}^2\end{equation}



(they can't both be less than half of $115797.28376308 \text{meters}^2/\text{s}^2$)


If inequality (2) , then
\begin{equation}s_{ub} \geq 57898.64188154\times 10^{13} \text{J}\end{equation}

This is clearly absurdly large.  To give you an idea of how large this is, this is about 6 times as energetic as the 2004 Indian Ocean Earthquake.  Obviously, $s_{ub}$, and hence s(r) is not just a quantum energy fluctuation.

On the other hand, if (3) is true
$S_q \geq 240.621 \text{meters \ s}$.

In human terms, this is a very large velocity, and is equal to 538.252 miles per hour.

Again, this is unacceptable.  Normal objects like cars and humans go much slower than this before turning around.  So, again, we arrive at an absurd physical requirement.

The only possibility, assuming s(r) is due to quantum fluctuations is that your initial premise, that objects always return under gravity is horribly flawed.

The ball is in your court.  Now, I spent a long time writing this up.  Let's see if you are honorable enough to respond accordingly rather than replying with a terse, irrelevant comment, like "it's just math!"

Thursday, September 11, 2014

An interesting way of looking at the Lorentz Transform

So, this is going to be a rather light posting today.  I was at first going to refer to it as short, but apparently, I can be quite verbose, so, so much for that.  Anyway, this post is really meant as a means for me to get the ball rolling on this blog again after a series of events that lead to me being unwilling to blog.  I actually hope to talk a little about that, at some point, but for now, here is a fun little thing I realized about the Lorentz Transform, which actually stemmed from my desire to find a way of assigning a coordinate to every point in space-time without explicitly relying on velocity.

Let me remind you of the Lorentz Transform, if you don't know what it is:
\begin{equation}
x' = \left (x - vt \right )\gamma
\end{equation}

\begin{equation}
y'=y
\end{equation}

\begin{equation}
z'=z
\end{equation}

\begin{equation}
t' = \left (t -\frac{xv}{c^2}\right )\gamma
\end{equation}

where
\begin{equation}
\gamma = \frac{1}{\sqrt{1-\frac{v^2}{c^2}}}
\end{equation}

If you aren't familiar with what this means, wikipedia has a good article on it.

And of course, the inverse transform is given by:
\begin{equation}
x = \left (x' + vt' \right )\gamma
\end{equation}

\begin{equation}
y=y'
\end{equation}

\begin{equation}
z=z'
\end{equation}

\begin{equation}
t = \left (t' +\frac{x'v}{c^2}\right )\gamma
\end{equation}

So, who cares?

Well, it occurred to me that I was wrong when I said that there was no way of uniquely identifying a space-time point without explicitly referring to velocity.  All it takes is a little adjustment to your point of view, and...well...cheating.

You see, you can actually use the inverse Lorentz Transform to fulfill this exact requirement, and the answer is quite trivial--express it in terms of the origin of the rest frame!  Let me just get into some philosophy before this.  When we describe the universe, at least in the non-general relativistic regime, it is always with respect to some fixed reference point.  Sure, you can express your system in polar coordinates, or spherical, or something completely exotic, but there will always be a point with the label (0,0,0,0).  Often, this point happens to be physically meaningful, and not just convenient.  It might represent some central object like the Sun, or the center of mass.  To show you the significance of this, let's try a thought experiment.

Pick a point--any point in the universe.  The only stipulation is that it must be uniquely defined.  Which did you choose?

I suspect that most people who perform this exercise would pick something that they could easily describe.  They might have chosen "my location", or "the center of Earth", or "the center of mass of Messier 83."  Some particularly creative people might have chosen something like "a point that is located 200 billion km in the direction of where the the geomagnetic field of Earth is pointing at my location."

But, here is what I suspect you didn't do--pick the center of mass of all discrete silicate glass materials in the visible universe with a charge of $-2\pm0.1$ coulombs.  Likewise, I doubt you picked the center of mass of the set of living organisms with the largest volume in the visible universe.

Now, you'll notice the difference between the first set of objects and the second.  In the first, each description, no matter how complicated, relied explicitly on some known physical reference point.  The objects in the second set did not explicitly reference any known physical point.  In fact, it is impossible, given our current knowledge of the universe, to determine what these points are.

We need some physical reference point to anchor our description of physical phenomena.  Even the second set of points, though not explicit, referenced a known location, as the visible universe represents the light cone of Earth, and that's before even talking about the need to express these things against the backdrop of an inertial reference frame to fully define them. (I would like to say, for the record, that this would be a fun, nerd game to play with all your nerd friends.  Specifically, see who can identify the most obscure point in space-time possible, especially with alcohol.  There is no doubt, that eventually, penises would work there way in there some where.  They always do.)

Thus, it seems that only by fixing some origin somewhere that we can begin to describe the universe.

If you think about it, this is exactly what we do with the inverse Galilean transform:
\begin{equation}
x = x' + vt'
\end{equation}

\begin{equation}
y = y'
\end{equation}

\begin{equation}
z = z'
\end{equation}

Namely, as long as everybody agrees where the reference point is, everybody can agree on where the event occurred, and since measurements approximate c as being infinite in speed, then everyone knows when it happened, too.

In fact, we can express equation 10 in another way.  Let $O$ represent the position of the reference point.

Then, we we have
\begin{equation}
x = x' - O
\end{equation}
Of course, this seems obvious.  All we're doing is describing points with respect to $O$.  However, to me, this represents an interesting way of looking at things.  In particular, by rephrasing the Galilean transform in this way, we no longer have to worry about actually measuring relative velocity.  All we need to do is figure out where $O$ was in relation to us, and we would be able to unambiguously identify a point in space that corresponds to every event.

So, the question is, can we use this as inspiration for the Lorentz Transform.  Of course!

So, take equation 6 and replace it with:
\begin{equation}
x = \left (x' - O \right )\gamma
\end{equation}
Now, this makes sense, at least partly.  Somebody moving relative to me at velocity $v$ would say that my trajectory is $-vt'$, assuming our origins coincided at when our clocks both read 0s.  Thus, if they treat me as their reference point, $O$, then this parallels the Galilean so far.
"But hold on!" you might protest, "$\gamma$ isn't just a constant.  It too depends on relative velocity."
Well, algebra's not a problem for us.  $O=-vt'$, so $v=-O/t'$, and this is where things begin to get really kind of interesting.  Let's see what this does to $\gamma$

\begin{equation}
\gamma = \frac{1}{\sqrt{1-\frac{\left( \frac{-O}{t'}\right)^2}{c^2}}}
\end{equation}
\begin{equation}
\gamma = \frac{1}{\sqrt{1-\frac{O^2}{(ct')^2}}}
\end{equation}

Now, notice that $ct'$ would be the position of an electromagnetic wave emitted along the x axis at the time the origins coincided.

So, $\gamma$ becomes an expression involving the position of O and the position of the electromagnetic wave.

We can do something similar with equation 9.
\begin{equation}
t = \left (t' -\frac{x'O}{t'c^2}\right )\gamma
\end{equation}
\begin{equation}
t = \left (\frac{(ct')^2 -x'O}{c(ct')}\right )\gamma
\end{equation}

Again, we see the equation clearly expressed in terms of the positions $O$ and $ct'$.

Now, see what I mean about cheating?  In some sense, this is a mathematical disappointment.  All I did was rewrite the Lorentz transformation, but in a physics sense, it has real ramifications for measurement--perhaps not profound--but real, nonetheless.  Just as with the inverse Galilean transform, we see that there is a way to uniquely assign a space-time coordinate to an event without having to measure velocity, by having everybody agree on a reference point.  In other words, you don't need to differentiate at all.  Just know where the reference point is at the time of the event, and whatever event you encounter, you can say exactly where it would have been for the person at rest with that reference point.  Keep in mind, also, that the fundamental forces are conservative, and therefore rely only position.  This form seems to respect that more than the conventional form.  In all honesty, though, I can't really say this is a useful way of looking at things.  But, then again, I have barely had any time to think about it.

So, what do you think?  Is there any use that can truly come from expressing the transform in this way?  I wonder if this might be useful as a pedagogical tool for talking about closed time-like curves resulting from FTL information transfer.  Anyway, I am signing off, for now.