The following articles are contained below:- A Brief History of Quantum Theory 1925 Y2K Fermions and Bosons; The Janus of Quantum Theory "GOD DOES NOT PLAY DICE WITH THE UNIVERSE" A. Einstein "EINSTEIN, DON'T TELL GOD WHAT TO DO" N. Bohr A Brief History of Quantum Theory 1925 Y2K Part I Inward Bound --- Sub atomic theory There are 4 formulations of modern quantum theory (QT). Historically it was Heisenbergs matrix mechanics, which was first to be published. The non-commutative nature of matrix multiplication is responsible for the appearance of his uncertainty principle. Schrodingers wave equation was quick to follow, having delayed publication due to his inability to produce a fully relativistic version. The nature of the wavepacket that describes a quantum entity, inherently produces the uncertainty in its position/momentum (the latter being related to the constituent wavelengths of the packet via de Broglies equation). Dirac showed the equivalence of both these methods and introduced his own formulation in terms of vectors in Hilbert space, using his <bra ket> notation. [Hilbert Space was originally utilized by von Neumann in his attempt to produce a universal axiomatic formulation for quantum theory]. Dirac was also the first to formulate a relativistic quantum theory, a union which demanded a field theory and necessitated the introduction of intrinsic spin and anti-matter! Finally there is Feynmans path integral (sum over histories) technique which is used in the more advanced treatment of interactions of particles with their fields (whose quanta are fermions and bosons respectively). If we consider a projectile such as a football, in its motion from the ground to its position of maximum height, its path is not that of the shortest route (i.e. a straight line) but rather (half) a parabolic curve. What is minimised however is its action i.e. its (kinetic potential) energy integrated over time, (this is known as the principle of least action). The fact that this so-called Lagrangian formulation is symmetric (unaltered) over local translations of space and time, yields the conservation of momentum and energy respectively. According to classical physics, objects move in a force field, so as to minimize this action and in just one page of his book, Dirac showed that the sum of the contributions of the various paths that a quanta can take (each with its own probability amplitude), reduces to that of the classical stationary action, in the limit that Plancks constant h tends to zero. In fact it is due to the small numerical value of h (~ 10^-34 Js) that quantum effects are not readily observed except at the atomic scale. In other words the principle of least action is due to quantum mechanics and the smallness of h. Feynman later elaborated this in his sum over histories technique, in which we associate a probability amplitude that is determined by the action of a particular path, and sum up all of these to find the overall probability of a quantum moving between two states (positions). This is schematically shown in the Feynman diagrams but it is important to realise, that the actual calculations involve the sum over all the possible paths and not just the particular lines shown in any given representative diagram. To first order effects, this technique just retrieves Schrodingers equation. However the power of this method is that we can utilize higher order perturbations to include the effect of movement (e.g. of an electron), in a field (electromagnetism), in a way which can also treat the field in a quantum mechanical manner (involving a photon), thus producing quantum electrodynamics (QED). [The concept of a force is therefore subsumed by that of an interaction and in the case of the weak interaction, there is no force manifested.] The path integral is also important because it makes quantum theories automatically consistent with special relativity, providing that the Lagrangian itself is already Lorentz covariant (i.e. it is not dependent on any one co-ordinate system of 4 dimensional space-time). Unfortunately these techniques are arduous, as indicated by the fact that it took 2 teams of mathematicians 4 years to work out one value (it took one year to discover an error in the calculation)! It did however result in 3 Nobel prizes but as Weinberg once protested "how can we make any money out of this". These perturbation techniques work by first using known exact solutions (which puts us in the right ball park), and then proceed by making more fine-tuned approximations. However at each progressively higher level of energy scale (corresponding to a greater level of complexity of Feynman diagrams), we need to renormalise our answers to ensure that all our probabilities again add up to unity. It is a bit like a fractal, in which approximate measurements of the perimeter of such a shape yields a different answer, depending on the magnification (energy scale) at which we carry out the measurement. Hence if left 'unregulated' these perturbation techniques are inherently divergent, producing infinities in the self energy of an electron. Renormalization is therefore a necessary technique of self consistent subtraction, which allows the theoretical values for the mass, charge and magnetic moment of an electron, to be in agreement with the experimental values (in particular those of the Lamb shift, the Lande factor g and the hyperfine anomaly). Much of the work up until 1950 was focused on QED, i.e. the production of a consistent quantum theory of the interaction of charged particles (electrons) with an electromagnetic field. Initial steps were undertaken by Dirac with his method of second quanization (an application of his beloved transformation theory), in which an electromagnetic field is represented as an assembly of of quanta (photons). This introduced creation and annihilation operators that obeyed certain commutational relations which ensured that the particles being described satisfied Bose-Einstein statistics. This field was not however relativistic (as is Maxwell's field description) but the method was extended to classical matter wave fields by Jordan, who introducing the correct anti-commutational rules, so that the particles (electrons), obeyed Fermi-Dirac statistics (but this again was not Lorentz covariant). [Jordan was the first to suggest that the commutational rules that effected the transition from the classical to the quantum description of a system of particles, also be applied to a system with an infinite number of degrees of freedom, that is to a field system.] Henceforth a quantum field could be interpreted as expressing the probability of finding a given distribution of particles states throughout space and time or alternatively an assembly of quanta could be described as a such a field, thus both the wave and particle nature are incorporated in the notion of a quantum field! Indeed the other great achievement of Dirac, was in producing the correct formulation for making quantum theory (QT) consistent with special relativity (SR) which also demonstrated the necessity of a field theory. The famous equation that bears his name, not only predicted the existence of antimatter and it also introduced the mathematical structure of spinor theory. Following on from all this endeavor, Heisenburg and Pauli developed a general method for quantizing (Maxwell's) fields, using the Lagrangian description and canonical variables (field potential and its conjugate momenta) but they encountered infinities when they tried to produce a complete picture of quantum electrodynamics (i.e. QED). The next major step was the independent work of Tomonaga and Schwinger who described the interaction of an electron matter field with the electromagnetic field but due to the effects such as vacuum polarization and the self energy of the electron, it became necessary to absorb these infinities at each level of the calculation. Schwinger achieved this by employing Dirac's contact transformation to eliminate such virtual effects in the calculation. A different approach was instigated by Feynman, who followed Dirac's philosophy in regarding electromagnetism as a quantized field but treated the electron as a quantum particle. In his path integral approach (which was also inspired by Dirac's work on the usefulness of the Lagrangian in QT), Feynman applied a relativistic cutoff field (which was equivalent to an auxiliary field) to cancel the infinite contributions due to the real particle of the original field. [The auxiliary masses of the cutoff are used more as a mathematical parameter, which finally tends to infinity and are nonobservable in principle]. Dirac however had strong reservations about any theory which neglects infinities instead of infinitesimals (as does calculus). However renormalization later became a guiding principle in the formation of the electroweak theory. Finally Dyson demonstrated the equivalence of Feynman's and Schwinger's method and did much to show that renormalization was consistent to all levels. [Schwingers method involved field operators and the calculations were very hard to follow and indeed had not been pursued beyond a second order, whereas Feynmans method had been more user friendly]. Dyson managed to recast the Schwinger formulation so that it displays all the advantages of the Feynman theory but unlike the path integral method, it used an electon-positron field and incorporated a term in the Hamiltonian, which described the interaction of this matter field with the Maxwell field. Although such approximation techniques are tedious it would be desirable to utilize Feynmans method for the other 3 interactions (i.e. the weak and strong nuclear forces and gravity). However unlike electromagnetism, where the strength with which charged particles couple to their field is small ~ 1/137, such perturbation techniques are more difficult with the strong interaction (Quantum Chromodynamics) and gravity has defied all attempts, since the calculations produce infinities that refuse to be renormalised. [Gluons interact with themselves as well as with their quarks, while gravitons, having energy and hence mass, also gravitate. Also, gravity is described as a curvature in space-time in general relativity (GR) hence a graviton implies a quantised space-time!!!] Interestingly it is Black holes which offer clues as to how to proceed with a theory of quantum gravity. The event horizon of such exotic entities can combine ideas from GR, thermodynamics and QT! The area of the event horizon can be related to temperature and hence entropy and has given rise to the development of what is known as the 'holographic principle'. This allows an explanation of what happens when the information (negentropy) stored in a (3-dimensional) assembly disappears into a black hole. Eventually due to Hawking radiation, the material but not the information, is restored to the universe at large, apparently in violation of the second law of thermodynamics. This dilemma is avoided if we assume that the information is stored in the 2-dimensional surface of the event horizon in the form of string segments (c.f. 'strings' below). Each minute segment of a string measuring 10E-35 centimeters across, functions as a bit. Hence the surface of a black hole is able to store 3-D information and is therefore analogous to a hologram. On another front, it would also be desirable to collect all the particles into a multiplet that becomes unified under one interaction, so that the laws of physics make no distinction between an electron, neutrino or a quark. Returning to the Lagrangian, certain local gauge (internal) symmetries, acting upon such a multiple particle state, implies the conservation of quantum properties but require the introduction of so-called gauge** fields, in order to ensure the invariance of the Lagrangian under such (local) transformations. [This involves Noether's theorem, in which symmetries are related to conservation laws; in particular local gauge symmetries demand the addition of a gauge field, in order that certain quantum values are conserved. It is this gauge field that is responsible for the observed fundamental interaction between the 'particles']. The interactions can therefore be viewed on a quantum level as gauge (Yang-Mills) fields, that must be introduced in order to ensure that the Lagrangian of (unified) particle states are symmetric under localizedinternal transformation. [The quanta of these gauge fields, being the photon, gluon and W, Z particles of the electromagnetic, strong and weak interactions respectively which are adequately explained by the 'standard model'. [Technicolour is a provisional theory which introduces a new interaction analogous to the colour force that binds quarks and attempts to go beyond the standard model but requires new generations of particles.] These are mathematically described by specific groups of transformations. Whereas electromagnetism and the weak nuclear interaction are successfully unified under a gauge group known as U(1)*SU(2), the strong nuclear force (Quantum Chromodynamics- QCD) is described by an SU(3) symmetry. We therefore need to find a master symmetry group, which subsumes these smaller symmetry gauge groups that are associated with some of these interactions. Regarding such superunification, it is not possible to satisfactorily incorporate gravity (which is governed by the non-compact Poincare group) in what is known as a Unitary representation (that dictates the other 3 quantum interactions), unless one resorts to supernumbers, which combines both fermions and bosons via supersymmetry. [Specifically there is a no-go theorem which states that a group that nontrivially combines both the Lorentz group and a compact Lie group cannot have a finite dimensional, unitary representation]. This concept was originally invoked in the early study of string theory but although this has received a recent surge in popularity, supersymmetry itself does not require a string formulation. [Such a Lagrangian, is that associated with the surface being swept out by a string, rather than that of a point like particle which, as it moves through time sweeps out a curve.] Indeed even without quantum theory, gravity can be regarded in terms of a local symmetry applied to special relativity and this puts gravity on an equal footing with acceleration, (where velocity and therefore the space-time frame of an observer, varies locally with its 4 dimensional position). We can therefore apply classical laws of physics to an accelerating frame of reference, if we invoke a gravitational field or alternatively a gravitational field can be eliminated locally by applying an accelerated frame of reference and hence the correct metric (curvature) of space-time that goes with it. A gravitational field therefore needs to be invoked, in order to allow the global space-time transformations of special relativity to be locally symmetric, thereby extending special relativity to that of general relativity. [In this respect gravity is said to be locally equivalent to acceleration.] Likewise the unified electroweak field needs to be invoked, if we are to allow a local gauge transformation to act symmetrically on a family of leptons or quarks. However whereas Feynmans path integral technique is applicable to these interacting particles, when considering gravity, it is not possible to use such accurate quantum techniques, since the calculations blow up into infinity and refuse to be renormalised. It is therefore hoped that by extending unification to include all 4 interactions some of these infinities will cancel out without the need for messy renormalisation tricks. So from the simple description of nature in terms of classical particles and fields, we have been forced to accept their description in terms of quantum theory. Group symmetry allows us a way of relating a collection of these particles, with the corresponding fields with which they interact with one another. Whereas Newton's gravitational law united falling bodies on Earth with celestial motion, Einstein unified space, time and gravity, while Maxwell succeeded in unifying Magnetism with Electricity. Gauge symmetry provides a way of combining both electromagnetism and the weak interaction and allows a way of extending this description so as to include the strong interaction (the so called standard model). However this effort has not achieved the status of a complete unification, since we have not yet discovered a single group, that encompasses all three interactions.(i.e. a Grand Unified Theory). Supersymmetry hopes to put the quanta of both particle and field(i.e. fermions and bosons), on an equal footing, while the latest development -- string theory, hopes to explain what symmetries are allowed, which in turn determine the conservation laws of physics. [Whereas bosonic fields behave in a conventional commuting manner, the fermionic field is a spinor representation of the Lorentz group and are consequently anticommuting!]. As an extra bonus, putting fermions and bosons in the same super multiparticle state, necessitates the introduction of a gravitational field (i.e. local supersymmetric transformations invoke a field that produces the localised space-time translations that is indicative of gravity). The price that we have to pay for such a simplification of physical phenomena into a unified framework, is ironically, a succession of layers of abstract concepts, together with their relevant mathematical structures. Hence from tangible atoms, we move by one level of abstraction to invisible fields and particles. A second level of abstraction takes us from fields and particles to the (gauge) symmetry-groups by which fields and particles are related. The third level of abstraction, is the interpretation of (super) symmetry-groups in terms of states in higher (10) dimensional space time, since it is the manner in which the space is compactified which determines the symmetries that are permitted. The fourth level is the world of superstrings by whose dynamic behavior the states are defined. Finally we arrive at M-theory, in which even the strings are regarded as just one of many possible p-branes that can exist in a 11-Dimensional space-time. M-Theory consequently contains many varied multiverses, of which our own particular universe permits such rich laws of physics, that it allows our very existence so that we can perceive it! * * * * * * * * * * * * * * * * * * * * * * Part II Outward Bound ----- Cosmology Quantum theory, has been used to apply a wavefunction to the universe as a whole (viz. The Wheeler De Witt equation), even though we are lacking a quantum theory of gravity. Such quantum cosmology has been used by Hawking et al, in order to help explain the ripples in the big bang radiation detected by COBE and to possibly explain how quantum wavefunctions can undergo decoherence to produce a classically observable universe. In the absence of a quantum theory of gravity, provisional Grand Unified Theories (GUTs) have provided an understanding as to why the universe is so smooth (homogeneous) and is so close to the critical density that would eventually halt its expansion, (referred to as the horizon and flatness problem respectively). In GUTs, the electroweak theory and the strong force QCD, are unified into one interaction and the underlying laws of physics make no distinction between an electron, neutrino or a quark. [In math speak, the SU(3)*SU(2)*U(1) gauge symmetry of the 'standard model' are unified under one larger symmetry group]. This symmetry between these particles (and their interactions) only becomes broken into the separate interactions that we observe around us today by a mechanism, which involves the so-called Grand Unified Higgs fields (different from the electroweak higgs field).. The spontaneous symmetry breaking (SSB) of the Higgs mechanism gives the particles their distinctive properties and their interactions different coupling strengths and ranges. In other words the distinction between strong, weak and electromagnetic interactions is caused by the way that their force carrying particles (i.e. gluons,W,Z and photons) interact with the different Higgs particles. The simplest GUT requires 24 Higgs fields but 2 Higgs fields can be represented as a 3 dimensional graph, which resembles a sombrero hat! It has the unusual property that its zero value has a (local) maximum energy density (centre of hat), while it achieves zero energy density at the rim, where the Higgs field is not at the origin (i.e. it is non zero), and it is here which corresponds to the SSB condition. When this SSB occurred in the very early universe, the Higgs fields aligns itself in a particular direction (analogous to a ball rolling down the sombrero to the brim in a spontaneously chosen direction) and the direction would be randomly chosen in different parts of the universe, which are sufficiently far apart. Such a mismatch in directions would correspond to a large number of magnetic monopoles being formed, which unlike electric monopoles (e.g. electrons) are not observed in our universe.[North and south poles occur together and are not observed separately, although Maxwells equations of electromagnetism would be more symmetric if we included such sources.] This is where Inflation(cf. addendum below) comes into play, since it allows us to alter the graph in a way that produces a dip (or a flattened top) in the centre of the sombrero which corresponds to a false vacuum. This false vacuum allows the alignment of the Higgs fields that are responsible for SSB, to grow gradually rather than produce a more chaotic distribution, which would result in the production of too many magnetic monopoles. As the name suggests, the false vacuum has peculiar properties, but the important point is that although it has a very high positive energy density, it has a far (~ 3 times) higher negative pressure and therefore corresponds to a repulsive force according to Einsteins general theory of relativity. [Note that it is not the uniform negative pressure that drives the expansion, since only pressure differences result in forces, but rather that pressure like mass, gives rise to gravitational forces and negative pressure gives rise to a negative repulsive gravity that is associated with a cosmological constant]. Hence when we have a SSB of a Higgs field coupled to gravity we generates a constant term, which corresponds an increase in the energy density of the vacuum which cannot be ignored when dealing with GR! Einstein originally introduced such a cosmological constant (lambda)** into his early model of the universe, which was then believed to be static, and hence the positive gravitational attraction of all the matter of the universe had to be balanced by a repulsive term. The introduction of such a term into the era when GUTs prevail, does however produce an exponentially inflating universe (the de Sitter solution), rather than the one we observe today that is typical of an initial explosion of matter, which is experiencing the decelerating effects of gravity. Incidentally the cosmological constant in the universe today is known to be close to zero, although quantum calculations involving GUTs produce an embarrassingly large value for lambda. Recent observations of supernova in distant galaxy clusters indicate that it has a small positive value, which consequently produces a repulsive 5th force, whose effects only become significant over large volumes of space. If these observations are indeed valid, it shows that the universe is undergoing acceleration and will never halt its expansion. The good news is that this explains the flatness of the universe without needing to invoke such a large value of the 'missing mass'. The larger the cosmological constant, the smaller he amount of positive mass/energy that is needed to produce a flat but not closed universe (Omega =1). Some modern notions have tried to do away with the somewhat ad hoc inflationary theory by allowing the value of lambda to decrease with time. The early universe thus underwent rapid expansion due to the large initial lambda value, which itself is not stable but decreases with time. This theory allows for a fine tuning, and as the universe expands and lambda decreases, matter is created in just the right amounts to keep the density of the universe sufficient for it to maintain its flat value. The problem with this theory is that it requires the universal speed of light to be coupled to lambda and decrease with time!! Such an iconoclastic notion as the speed of light being reduced as the universe expands, will require a good deal of empirical evidence that is not easy to produce.(C.f.** below). An alternative approach is to introduce another quantum field called Quintessence, which is a basic (more adaptive) property of 'empty' space that may be able to account for the 5th force of repulsion. So historically inflation was used to explain the dearth of magnetic monopoles. However by incorporating such a brief but rapid expansion in the early universe, we can also allow for our observable universe to have been causally connected in the past and these effects can explain the flatness and the horizon problem respectively. The false vacuums allows the Higgs fields to grow in alignment without too much chaos, therefore avoiding the creation of too many magnetic monopoles, before they quickly decay into the true vacuum (via quantum funneling) that we observe today. Unfortunately, calculations involving the growth and percolation of these false vacuums have encountered difficulties and several refined variations of inflation theory have had to be introduced. In addition, if we resort to string theory, it needs to be embedded in 26 dimensions that became compactified during the early SSB period, when the false vacuum collapsed (as if things arent complicated enough). It therefore becomes necessary to study the actual space in which the field equations exist. Whereas the mathematics of group theory was required by the physicists of the 60s, topology has also become necessary for those studying how higher dimensional universes become curled up to produce our observed 4 dimensions of space-time. Whats known as the cohomolgy properties of these compactified spaces, are related to the number of generations (viz 3) of quarks and leptons that can occur, as well as the symmetries that are permitted (this number equals half the Euler characteristic of the manifold). This prompts a small digression, in that Penrose employs a sheaf version of cohomology in his twistor** theory. His philosophy is that both quantum theory and general relativity need to be replaced by a single and more fundamental theory, one which would replace the space-time description of physics, with that of a space of complex numbers associated with spinor mathematics projective twistor space. Spacetime and twister space are related by a correspondence that represents light rays in spacetime as points in twister space. A point in spacetime is then represented by the set of light rays that pass through it (each having a 'null direction'). Thus a point in spacetime becomes a Reimann sphere (a well known method of steriographically representing a complex plane) in twister space and the effect of Lorentz transformations produce congruences that twist (hence the name). In this way we consider spacetime to be a secondary concept and regard twister space (initially the space of light rays) to be a more fundamental space. The actual twister coordinate, is constructed out of a pair of two-spinors which obey certain relationships to each other, that are dictated by the corresponding spacetime coordinate (the spinors satisfy a specific twistor differential equation). Now if we think of light rays as photon histories, we also need to take into account their energy and also its helicity which can be either right or left handed or zero. The multiplication of a twister by its conjugate yields twice the value of its helicy or twist, which can therefore be either positive , negative or zero. [This then leads to the formation of a projected twistor space, in which the full twistor space is projected into a space which is divided into the three regions according to whether their helicy is positive, negative or zero].Twister theory is a conformal (i.e. scale invariant ) theory, in which physical laws that are usually written in the framework of space-time, are expressed in what may turn out to be a more fundamental and revealing manifold, which could therefore yield a more fundamental understanding of physics. A massless field is then defined by a contour integral in twistor space and these integrals are determined once we know about the singularities of a general twistor function in twistor space (in the case of electric or magnetic fields these singularities look like charges or sources where field lines begin or end). In other words, the differential equations that describe fields in space time have been reduced to simple functions in the geometry of complex spaces, viz. projective twistor space. The field is given by the contour integral around this function which in turn is determined by the nature of the singularity of this function, (hence the use of sheaf cohomology in dealing with the analysis of these regions). For massless fields, it turns out that the helicity of the quantum particle H is related to the homogeneity of the twister field T by the simple relationship; H=h/2 (-2-(T)), so if we have a photon of spin +1 it is necessary to write down a twister function of general power of 4 while for a -1 helicity photon the homogeneity must be 0. (for a H=+2 graviton, T=6). We therefore have a fundamental distinction between left and right handed light i.e. chiral asymmetry is a fundamental property of twister theory! Producing a quantum theory of gravity is particularly difficult for several reasons. Firstly quantum gravity implies quantised space time!!! Secondly gravity is intrinsically so weak it has not yet been possible to detect classical gravitational waves, let alone its graviton (it is also difficult to combine the two realms, since QT deals with very small scale phenomena, while GR deals with large masses). Thirdly, quantum field theories are written as spinor fields, to which, the Riemannian techniques of GR not applicable Finally regarding unification, the GR group structure is non compact, while QT group is Unitary and there is a theorem which states that there is no finite unitary representation of a non compact group (this is why we have to resort to supernumbers). At first glance GR and QT look very differently mathematically, as one deals with space-time and direct observables while the other with Hilbert space and operators. One approach to synthesizing the two and providing a quantum theory of gravity, involves Topological Quantum Field Theories (TQFT). [A topologist is sometimes defined as a mathematician who cannot tell the difference between a tea cup and a doughnut, since they are diffeomorphic to each other, both having a genus of 1 ]. Quantum states are given topologies and cobordism allows a description of how quantum (gravity) states evolve i.e. TQFT maps structures in differential topology to corresponding structures in quantum theory. The state of the universe can only change when the topology of space itself changes and TQFT does not therefore presume a given fixed topology for space-time. Quantum operators are therefore related to cobordism and n-category theory (i.e. algebra of n-dimensions) is a useful advance in understanding the cobordism theory of TQFT. There have been two reactions to the apparent inconsistency of quantum theories with the necessary background-independence of general relativity. The first is that the geometric interpretation of GR is not fundamental, but just an emergent quality of some background-dependent theory. The opposing view is that background-independence is fundamental, and quantum mechanics needs to be generalized to settings where there is no a-priori specified space-time. This geometric point of view is the one expounded in TQFT. In recent years, progress has been rapid on both fronts, leading ultimately to String Theory (which is not background independent) and Loop Quantum Gravity (LQG), which is background independent and also incorporates the diffeomorphic invariance of GR. Topological quantum field theory provided an example of background-independent quantum theory, but with no local degrees of freedom, and only finitely many degrees of freedom globally. This is inadequate to describe gravity, since even in the vacuum, the metric has local degrees of freedom according to general relativity (e.g. those due to the propagation of gravity waves in empty space). Loop Quantum Gravity is a nonperturbative quantization of 3-space geometry, with quantized area and volume operators. In LQG, the fabric of space-time is a foamy network of interacting loops mathematically described by spin networks (an evolving spin network is termed a spin foam; spin foams are to operators what spin networks are to states/bases). These loops are about 10E-35 meters in size, called the Planck scale. In previous lattice theories the field is represented by quantised tubes/strings of flux which only exist on the edges of the lattice and the field strength is given by the value of integrating around a closed loop. In LQG space and time are relational! As in GR where there are many ways of slicing a section of space time, there are many ways of slicing an evolving spin network - thus there are no things only processes! [A spin network is a graph with edges labeled by representations of some group and vertices labeled by intertwining operators. Thanks in part to the introduction of spin network techniques, we now have a mathematically rigorous and intuitively compelling picture of the kinematical aspects of loop quantum gravity.] The loops knot together forming edges, surfaces, and vertices, much as do soap bubbles joined together. In other words, space-time itself is quantized. Any attempt to divide a loop would, if successful, cause it to divide into two loops each with the original size. In LQG, spin networks represent the quantum states of the geometry of relative space-time. Looked at another way, Einstein's theory of general relativity is a classical approximation of a quantized geometry. Historically, Regge calculus was the first attempt to quantize gravity, by dividing up space into small (flat) 3D tetrahedral simplexes, in which the curvature is concentrated along their boundaries. From this Hilbert space of quantum tetrahedron, it was intended to produce 'Feynman propagators' for gravity and recover Einstein's field equation in the macroscopic domain of space-time. Each of the edges is associated a spin j and in one type of approach, the exponential of the action in such a configuration, is a suitable product of the the 6j symbols associated to each of the 6 edges of the 3-simplexes and a partition function is obtained by taking the sum of this products over the possible associations of the spin to the edges. Remember that angular momentum is a quantum (bi)vector and is therefore subject to Heisenberg' Uncertainty Principle. [In considering a Lorentzian (as opposed to a 3-D) spin network we employ a 4-simplex in which there are 4 tetrahedra, 5 vertices, 10 triangles and 10J symbols]. Next consider Wilson loops, which have been used to analyze fields in QCD by means of applying a lattice structure and integrating along closed paths. They are functions on the space of connections; at a lattice point the Wilson loop is just the trace of holonomy around the loop on the lattice, taken in some representation of the holonomic group of the gauge field. [This philosophy originated from considering the vacuum as being like the discrete lines of (magnetic) flux that is exhibited by superconductors.] A 'connection' on the lattice is simply an assignment of an element of a gauge group to each edge of the graph, representing the effect of a parallel transport along the edge. [The holonomy around such a gauge field, is a measure of the field strength, which in turn determines the value of the Feynman path integral]. Each edge of this lattice is assigned a gauge group element that represents the (holonomic) connection and the vertex is also assigned a group element that represents a gauge transformation. From this a quotient space is formed (i.e. the space of connections modulo gauge transformations), and by mapping these onto suitable(complex) irreducible spin representations, we obtain a suitable way of producing a Spin Network. [Such spin network edges represent quantized flux lines of the field]. Holonomy is a natural variable in a Yang-Mills gauge theory, in which the relevant variables do not refer to what happened at a point but rather refers to the relation between different points connected by a line (curve). Hence to create a framework for quantum GR we introduce a connection A and a momentum conjugate E, from which we can produce a spin network together with area and volume operators, which act upon space so as to quantise it. [The flux through a surface is represented by area operators for a spin network, acting on a surface described by a spin network basis]. In LQG, holonomy becomes a quantum operator that creates loop states. Over a continuous background, Wilson loop states are far too big to produce a basis of Hilbert space of a QFT. However loop states are not to singular or to many when formulated in a background inependent theory, where spacetime is itself formed by loop states, since the position of these states is relative only with respect to other loops and not to the backlground. Therefore the size of the space is dramatically reduced by this diffeomorphic invariance (a feature of GR itself). A finite linear combination of loop states are defined precisely as spin network states of a Yang-Mills theory. Spin networks are gauge invariant and by taking suitable sums of tensor products, provide an orthonormal basis for LQG. Penrose had earlier introduced Spin Networks, in which the edges were labeled by an irreducible representation of a SU(2) Lie gauge group (characterized only by its dimension d =2j+1, where j is the quantum spin number) and the vertices with intertwining operators (tensors that transform incoming states into the outgoing states) and it was found that such a combinatorial formalism was preferable since it produced a relational theory. Spin networks have edges which are associated with a spin j quantum of action and the number of edges intersecting at a node determines its area, while the volume is determined by the number of nodes in a given region. A strict connection exists between quantum tetrahedra and 4-valent vertices of SU(2) spin networks. The 4 faces of a tetrahedron are associated with 4 irreducible representations of SU(2), which are represented by a perpendicular line, the 4 of which meet at a central node of the tetrahedron (there are actually bivectors associated with each face, in keeping with constraints of GR relating to the Ricci curvature tensor --- quantizing the bivectors/tetrahedra amounts to labeling each face with a pair of spinors). We therefore obtain a 4-valent (colour coded) spin network (each line of which represents a quantized unit of action, while the nodes behave as area operators), which can exhibit properties that are gauge invariant. The quantum bivectors allow us to construct area and volume operators which act upon the spin network basis to produce a discrete spectrum which has units of Planck length squared and cubed respectively (in the case of the area operator, the eigenvalues are 1/2sqrt (j(j+1)) and as in keeping with quantum theory we would expect these to correspond to physical observable i.e. we have a quantized space! These spin networks (formed from the above mentioned quotient gauge space), do not refer to a specific space background and we can reproduce Wilson loop calculation to imitate a quantum theory of gravity, which is relational, as in the spirit of GR. So, since spin networks form a convenient basis of kinematical states, they have largely replaced collections of loops as our basic model for 'quantum 3-geometries'. Now in order to better understand the dynamical aspects of quantum gravity, we would also like a model for 'quantum 4-geometries'. In other words, we want a truly quantum-mechanical description of the geometry of spacetime. Recently the notion of 'spin foam' has emerged as an interesting candidate; so whereas spin networks provide a language for describing the quantum geometry of space, a spin foam attempts to extend this language to describe the quantum geometry of space-time. A spin foam is a 2-dimensional cell complex with faces labeled by representations and edges labeled by intertwining operators; generically, any slice of a spin foam gives a spin network. We calculate the amplitude for any spin foam, as a product of the face and edge amplitudes (which equate to propagators) and the vertex amplitudes (which equate to intersection amplitudes). [Abstractly, a Feynman diagram can be thought of as a graph with edges 'labelled' by a group representation and vertices labelled by intertwining operators. Feynman diagrams are 1D because they describe particle histories, while spin foams are 2D because in LQG, the gravitational field is described not in terms of point particles but as 1D spin networks. Feynman computes probability in terms of probability amplitudes for edges and vertices whereas spin foams compute probability amplitudes as a product of faces, edges, and vertices amplitudes. Like Feynman diagrams spin foams serve as a basis for quantum histories.] Although QLG has been successful in predicting Hawking radiation and
Black Hole entropy, it is restricted to the domain of quantum gravity
and as yet does not offer any import on the other fundamental interactions
or the possibility of unification. Unlike string theory it does however
offer testable predictions, such as the variation of the speed of light
at different energies. The spin foam which makes up the fabric of space-time
predicts a varying refraction coefficient depending on the frequency and
hence energy of the photon. It is therefore hoped that by studying gamma
ray bursts from the most remote regions of the universe, this small dispersion
in the arrival times of the radiation can be observed (other avenues of
research also point to theories involving a variable speed of light).
Some researchers believe that even the success of string theory can be
explained in terms of discrete units of space that become evident on the
Planck scale (which being ~ 10 E -35m, is much smaller than that of the
compactified dimensions of superstring theory). Also both theories allude
to a version of the Holographic**
principle (where a bulk theory with gravity in n dimensions, is equivalent
to a boundary theory in n-1 dimensions without gravity, cf. Maldacina
conjecture below), in which entities such as black hole contain all their
information in their Event Horizons - one bit for every 4 Planck areas.
This arises since black holes emit Hawking radiation and therefore its
mass is related to a thermodynamic temperature, hence the entropy of a
black hole is proportional to its surface area while information is negentropy.
However LQG emphasize the necessity to have a relational theory
in which space and time are dynamic rather than fixed and the primary
concept is that of processes by which information is conveyed from one
part of the world to another. Hence the area of any surface in (QLG) space
is nothing but the capacity of that surface as a channel of information
and so the world can be construed as a network of relationships. This
avenue of approach has lead to the study of a relational logic called
Topos theory and non commutative geometry, in which it is impossible
to determine enough information to locate a point (a point is then described
by an appropriate matrix) but it can support a description of particles
and fields evolving in time) Hawking on the other hand believes that QT of supergravity is the way forward, although he does utilize an imaginary time coordinate. Time has been a problem for philosophers and more recently, physicists have started to take the view that our concept of time may be responsible for some of the intractable problems that face QT and GR (particularly relating to the collapse of the wavefunction but also a QT of gravity implies quantized space-time). Indeed J. Barbour believes that time does not exist but is merely a mental construct. He has a tentative theory which involves a stationary cosmological wave functions (akin to that used in the Wheeler- De Witt equation) acting upon a configuration space of the whole universe (known as Platonia). Einstein once said that space and time are modes by which we think rather than conditions in which we live, but as Hume would say how can something that exists as a series of states (the Nows) be aware of itself as a series? Of course in the cosmic wavefunction, we must also include the human mind but I am still not convinced that this will be sufficient to completely justify Barbours claim of The End of Time in physics. Returning to string theory, five versions originally evolved but more recently, a second revolution has subjugated these into an 11 dimensional formulation known as M-theory which also accommodates supergravity! (M stands for mystery, as very little is known about the underlying nature of this 21st Century theory). As T.S. Eliot once wrote, "trust the tale not the teller", hence even if the models that are proposed appear incredulous, the mathematics may hold real value. To summarise therefore, inflation theory is good example of where studies in sub-atomic physics are relevant to those of cosmology. When we study the quantum mechanics of GUT and add on the effects of general relativity as an appendage, we find that it necessary to include a non-zero value for the cosmological constant for the early universe, when strong, weak and electromagnetic interactions were unified. This produces an exponentially inflating universe during this early phase transition, which overcomes the horizon, and flatness problems. Its main raison dêtre was however to explain the scarcity of magnetic monopoles. These would otherwise be more prevalent if there had been a more chaotic Higgs mechanism, rather than the inflationary version of the SSB that is responsible for the distinct interactions that we observe today. The richness of these physical laws is nowhere more evident, than in the creation of our own planet Earth and in particular, the evolution of the human mind, which allows contemplation of the very laws that are responsible for its existence.
Addendum ** In some sense complex numbers are more fundamental than real numbers. The extra mathematical structure gives rise to theorems like Cauchy's integral formula, which demonstrates an interplay between local and global (topological) properties of a manifold and this has been exploited to great effect in twistor theory. [This could have relevance to the non-locality involved in the actual collapse of a wavefunction or the fact that the negative energy stored in an gravitational field is also non local] There are also local isomorphisms between space-time groups and those groups associated with complex spaces (the latter, as in the case of spinor space, is simply connected, while the rotational group of Euclidean geometry is doubly connected). Indeed quantum mechanics demands a complex space (that of spinors) in order for it to be compatible with SR, this being part and parcel of quantum field theory. Now a spin n/2 field, which is represented by a symmetrical spinor with n indices, can be associated (in Projective Twistor space), with a twistor function f{-n-2} of homogeneity {-n-2} via a contour integral. One of the most natural ways in which complex analysis and contour integral techniques can be developed, is to consider sheaf theory and sheaf cohomology, which takes a fundamental role in twistor theory. More precisely, zero rest mass fields on space-time, which are originally described as real spinor fields satisfying field equations, can be expressed as arbitrary holomorphic functions of a twistor (with contour), and these can then be further interpreted as elements of a (first) sheaf cohomology group. [Historically, it was the study of sheaf theory, which originally gave rise to CATEGORY theory, in an attempt to determine which given sheafs are equivalent. For example, a presheaf of abelian groups over a topological space X, is a contravariant functor on the category of open subsets of X] **The cosmological constant Lambda acts like a (positive) vacuum energy density but is associated with a negative pressure because; 1. The vacuum energy density must be constant because there is nothing for it to depend upon. 2. If a piston capping a cylinder of vacuum, is pulled out producing more vacuum, the vacuum within the cylinder has more energy, which must have been supplied by the force pulling on the piston. 3. If the vacuum is trying to pull the piston back into the cylinder, it must have a negative pressure, since a positive pressure would tend to push the piston out. If we denote the ratio of the vacuum density/critical density as Kappa and the ratio of the actual matter density /critical density as Omega, then obviously the Universe is open if Omega + Kappa is less than unity, closed if it is greater than unity and flat if it is exactly one. Now if Kappa is greater than unity then the Universe will expand forever unless Omega is much larger than current observations suggest. For Kappa greater than zero even a closed Universe can expand forever (e.g if the density of matter is as small as we currently observe).
MORE DETAILS REGARDING INFLATION The electroweak Higgs field, is responsible
for spontaneous symmetry breaking and gives mass to particles such as
the W and Z particle. In supersymmetry theory it is believed to be responsible
for giving mass to all the particles, depending on how strong it couples
to each of them. The Grand Unified Higgs mechanism may also explain the
nature of the rapid inflation experienced during the earliest epoch of
the universe, since its initially high energy value can be associated
with a negative pressure which drives the expansion. When the energy of
the vacuum drops to zero (being converted into matter and radiation),
the Higgs field takes on a non zero value, which breaks the symmetry of
the known interactions by giving mass to the particles and their force
carrying bosons. This inflation field solves the cosmological problems
of flatness, monopole scarcity and the horizon problem. and was derived
from a consideration of the adding Einsteins field equation to GUT.
During this inflationary epoch which lasted for only a billionth of a
a billionth of a a billionth of a second a region the size of a DNA molecule
would have expanded to the size of the Milky Way galaxy - - a greater
expansion than that that has occurred during the remaining 13.7 billion
years of the universe. Only by vaporizing the vacuum by returning
it to a temperature high enough to evaporate the Higgs field (causing
it to have a zero average value throughout space), would the full symmetry/unification
become apparent again. According to theory, matter and radiation are produced at the end of inflation and however, whereas these decrease in energy as the universe expands, the inflationary field actually increases its energy (sapping energy from the expansion like a rubber sheet that is expanding), since it initially has a constant density as it expands. As the universe expands, matter and radiation lose energy to gravity, while an inflation field gains energy from 'repulsive' gravity. In this way we can account for the large amount of matter that is present in the universe today, by just considering a small amount of matter (~20 pounds) that originated in a tiny space (10^-26cm diameter), as the inflationary phase grew to a close. This is because at the onset of inflation, the field didn't need to have much energy, since the enormous expansion it was about to sporn would enormously amplify the energy it carried. [This is in contrast with the standard big bang model in which the initial energy content of the universe would need to be enormous, due to its high energy content] hence inflationary cosmology, by 'mining' gravity can produce all the ordinary matter and radiation that we see today from a tiny speck of inflationary filled space. Note also that small quantum fluctuations in the early inflationary Higgs field arise due to Heisenbergs uncertainty principle and these can explain the seeding of the universe with inhomogeneities that coalesced to form present day galaxies. This theory is in close agreement with recent analysis of Cosmic Background Radiation. Inflation also rescues the problem regarding why there is a low entropy in the early universe, which gives rise to 'the arrow of time'. During the inflation epoch the residing field made gravity repulsive, which meant that any clumpiness in the fabric of the cosmos (due to Heisenberg's UP), was expanded so rapidly and to such an extent, that it these warps became stretched quite smooth. Also because of this rapid expansion of space the density of matter also became very diluted. This is in direct contrast to the post inflationary period where the attractive force of gravity would tend to cause the creases of space and clumping of matter to increase. At the end of the inflationary period, the universe would have grown enormously removing most of the nonconformitiey, and as the field reached the bottom of the potential energy well, it converted its pent up energy into a nearly uniform bath of particles. Hence this smooth uniform spatial expansion of matter that results in the inflation model explains why the universe started with a relatively low entropy state. In actual fact the overall entropy increases during inflation (in keeping with the second law of thermodynamics), but much less than it might have been, since the smoothing has reduced the gravitational contribution to entropy (uniformity and reduced clumping means less entropy -- black holes being the most extreme case of clumping having the greatest entropy). When the inflationary field relinquished its energy it is estimated to have created 10^80 particles of matter and radiation and this would correspond to an enormous increase in entropy. So even though the gravitational entropy went down this was more than compensated by the creation of more particle states in accordance with the second law of thermodynamics. The important thing is however that inflationary expansion/smoothing of the universe created a huge gap between what the entropy contribution of the universe was and might have been, thus producing a low entropy in the early universe. This then set the stage for a billion years of gravitational collapse (increasing entropy), whose effects produced the seeding of galaxies and the formation of stars that we continue to witness, as the arrow of time moves forward. In the above considerations it is necessary to assume that prior to inflation there was a typical high entropy space which was riddled with warps and that the inflationary field was highly disordered with its value jumping around due to quantum uncertainties. [Remember that the UP relates to the complementary uncertainties in a particles position and momentum (~ rate of change of positioned) and when applied to fields, it implies that the more we know about the value of a field (number of particles at a location),.the less we know about its rate of change at that location]. Hence we can expect that sooner or later a chance fluctuation within this turbulent highly energetic preinflationary period, will cause the value of its field to jump to the correct value in some minutel region of space, and in so doing initiate an outward burst of inflationary expansion. Now this understanding is actually close to Boltzmann's suggestion, that what we now see is a result of a rare but every so often expected fluctuation from total disorder. The advantage here however, is that only a small fluctuation within a tiny space during the inflationary period, is necessary to yield a huge and highly ordered universe that we now observe. A jump to lower entropy within a tiny space was leveraged by inflationary expansion into the vast reaches of the cosmos. Thus inflation doesn't just explain the horizon problem and the dearth of magnetic monopoles, it also accounts for the low entropy of the early universe (arrow of time) the flatness of space and the slight inhomogeneities from which galaxies formed. Indeed it is also conjectured that such a sprouting of an inflationary universe could be common enough to occurred elsewhere during the chaotic primordial state of high entropy and may continue to do so, repeatedly sprouting new universes from older ones, creating a multiverse (each maybe with different subsequent laws of physics which evolved at various symmetry breaking phases). Note that in order to artificial create conditions for such a universe, we would have to cram about 20 pounds of inflationary field into a space of about 10^-78 centimeters cubed which means a density of 10^67 times that of an atomic nucleus. In summary, through a chance but every so often expected fluctuation of a chaotic (high entropy) primordial state, a tiny region of space achieved conditions that lead to a brief burst of inflationary expansion. The enormous rapid expansion resulted in space being stretched tremendously large and extremely smooth. As this burst drew to an end, the inflationary field relinquished its huge amplified energy by filling space nearly uniformly with matter and radiation. As the inflation's repulsive gravity diminished, ordinary attractive gravity became dominant. Gravity exploits tiny inhomogeneities caused by quantum fluctuations that are inherent in Heisenberg's UP, and causes matter to clump, forming galaxies and stars and ultimately leading to the formation of solar systems such as our own. After about 7 billion years repulsive gravity once again became dominant but this is only relevant on largest of cosmic scales and the nature of this dark energy and indeed the known dark mater is as yet unknown. Footnote: . . There are still many fine tuning problems with inflation, and as such inflationary cosmology is not a unique theory and many different versions have been proposed. These include ; Old inflation, new inflation, warm inflation, hybrid inflation, hyper inflation, assisted inflation, eternal inflation, extended inflation, chaotic inflation, double inflation, weak inflation, hypernatural inflation
**The historic sequence for this is as follows; Hawkings originally showed that the area of a black hole can never decrease
in time, suggesting that it is analogous to entropy. Bekenstein argued
that this relationship was more than just an an analogy and that a black
hole has an actual entropy that is proportional to its area. However entropy
is related to temperature (~S = ~Q/T)and this was not a property that
was associated to black holes at that time. Hawkings then went on to discover
that a black hole emits radiation and can be given a temperature that
is inversely proportional to its mass (and the mass of a black hole directly
determines its event horizon). Hence the entropy of a black hole can be
shown to be S =1A/4hG and when object fall into a black hole the entropy
of the surrounding universe decreases (since negentropy is a measure of
information which is reduced when objects disappear,) while that of the
black hole increases so as to maintain/increase the overall value. On
the other hand if the black hole radiates energy, it loses surface area
and hence entropy but the entropy of the outside world will increase to
make up for it. However the entropy that is removed when a highly organized
system is dropped into a black hole, is much less that the increase of
entropy that is returned, when the black hole radiates that amount of
mass back to the universe, thus implying an overall increase in entropy,
in keeping with the second law of thermodynamics. Bekenstein later went
on to assert that the amount of information that can be contained in a
region is not only finite but is proportional to the area bounded
by that region, measured in Planck units (Holographic principle) and this
implies that the universe must be discrete on the Planck scale. This Bekenstein
Bound is partly a consequence of GR and the 2nd law of Thermodynamics,
but the argument can be turned around, and it can be shown that assuming
the 2nd law and the Bekenstein Bound, it is possible to derive GR. Hence
we have 3 approaches to combining GR with QT viz String theory, LQG, and
black hole thermodynamics and each of these indicate (in differing ways)
that space and time are discrete (the last two are also relational based).
More recently, new difficulties have become evident, from the fact that
when (organized) objects drop into a black hole, their quantum wave functions
are in a pure (correlated) state, while when the Black Hole eventually
evaporates, the radiation is in a mixed quantum state (i.e. the individual
quanta are not correlated as in an assembly of bosons or fermions). Now
a pure state cannot evolve into a mixed state by means of
a unitary transformation, which is a problem since unitary transformations
are a crucial feature of all quantum wavefunctions (in order that probabilities
evolve in a correct manner). Hence we need to find a way of reconciling
this dilemma, perhaps by invoking a non unitary theory or by discovering
a way of accounting for the extra information that a pure state has in
comparison to a mixed state. If the correlations between the inside and
outside of the black hole are not restored during the evaporation process,
then by the time that the black hole has evaporated completely, an initial
pure state will have evolved to a mixed state, i.e., "information"
will have been lost. For this reason, the issue of whether a pure state
can evolve to a mixed state in the process of black hole formation and
evaporation is usually referred to as the "black hole information
paradox". [There are in fact two logically independent grounds for
the claim that the evolution of an initial pure state to a final mixed
state is in confict with quantum mechanics: Some advocate that a black hole cannot completely disappear but that some of the original information persists as a kind of nugget. Others believe that this information is re-emitted as some form of correlated particles. Another view is that a new quantum theory of gravity will necessarily be non unitary. Hawking has changed his original view that information is lost and now advocate that the actual probability of sub atomic(and virtual) black holes actually causing the loss of information is minuscule and that unitarity is only violated in a mild sense. [He proposes that in the presence of black holes, the quantum state of a system evolves into a (non-pure state) density matrix --- a la von Neumann] This is somewhat analogous to the improbable effect in the violation of the 2nd law of thermodynamics or the approach to decoherence in QT. In effect the sum over all the possible histories (geometries) of the universe, results in the nullification of the non unitary effect of black holes in the long term. Susskind on the other hand has applied t' Hooft's holographic principle to sting theory and believes that the information is stored in the horizon of a black hole [Indeed using string theory to calculate the possible configurations of black holes has reproduced the BH formula for its entropy -- as have LQG calculations]. This has been strengthened more recently by Maldacinas conjecture (AdS/CFT correspondence), which demonstrates that there is an equivalence between the string theory of gravity in a 5D anti-de Sitter universe and the conformal supersymmetric Yang- Mills theory on its horizon [This holographic duality becomes more precise the larger the value of supersymmetry N]. Maldacina's conjecture not only says that gravity is in some deep way the same as quantum field theory but also implements the holographic principle in a concrete way. Proof of Bekenstein Bound Assume the contrary, that there is an entity 'A' such that the amount of information needed to describe it, is much larger than its surface area. Now 'A' cannot be a black hole because we know that the entropy of a black holes that can fit inside 'A' must be equivalent to an area less than that of 'A' but in this case its entropy must be lower than the area of the screen in Planck units. If we assume that the entropy of the black hole counts the number of its possible quantum states, this is much less than the information contained in 'A'. Now 'A', according to GR has less energy than a black hole that just fits inside 'A' and so as we add energy to 'A' we reach a point at which it will collapse into a black hole. However it now has an entropy which is equal to a quarter of the area of 'A'. Since this is lower than the entropy of the original 'A' we have succeeded in lowering the entropy of a system, which contradicts the second law of thermodynamics! Consequently if we accept the 2nd law of thermodynamics we must believe that the most entropy that we can attribute to 'A' is a quarter of its area Bekenstein's law. With every horizon that forms a boundary separating an observer from a region which is hidden from them, there is associated an entropy which measures the amount of information which is hidden behind it (and is proportional to the area of the horizon) Unruh's law. Accelerating observers see themselves as embedded in a gas of hot photons (Rindler particles) at a temperature proportional to their acceleration. {As their acceleration increases they perceive more virtual radiation. Now although this radiation is (non-locally) correlated, their observable horizon shrinks. This is because light from a sufficientl distant source can never catch up with a continually accelerating observer, even though they can never reach the speed of light -- the greater the acceleration the nearer is this horizon . Hence because much of their world cannot be seen, this produces a randomness which implies an increased temperature. This is in accordance with the principle of equivalence; for both black holes and accelerating observers, the smaller the 'horizon' the greater the random radiation (temperature). However in the case of the former this area is the region into which information has fallen and increases with entropy, while in the later, the visible region that surrounds them reduces with increased acceleration, meaning that the universe of information that is inaccessible increases. In both instances, there is an increased temperature and decreased entropy, as the horizon is reduced . . . deltaS = deltaQ/T } * * * * * * * * NB. . The entropy of a black hole is proportional to the area of its event horizon and has an enormous value. [Entropy can be regarded as a measure of how probability that system is to come about by chance] Now because of the large black holes at the centres of galaxies, this black hole entropy easily represents the major contribution to the entropy of the known universe! Because of the second law of thermodynamics, the entropy at the beginning of the universe must be very small. This beginning is represented by the singularity of the big bang, which must be extraordinarily special or fine tuned compared to the high entropy singularities that lie at the centre of black holes. It therefore requires a new theory -- quantum gravity -- to explain the remarkable time asymmetry between the singularities in black holes and that at the big bang.
Fermions and Bosons; The Janus of Quantum Theory In Newtonian mechanics objects are described classically as particles in which the position and momentum at a particular instant can be specified exactly. A collection of particles in thermal equilibrium is described in terms of Boltzmann statistics, in which particles are distinguishable. In quantum mechanics a 'particle' is described by a wavefunction which is acted upon by hermitian operators which extracts eigenvalues that represent measurable observables. Such observables as momentum and position although complimentary are also mutually exclusive (as enshrined by Heisenbergs uncertainty principle). A wavefunction does however have an internal structure known as a spinor which is crucial in determining how the particle behaves. A collection of such quanta is described either by Einstein-Bose statistics (in the case of integer spin bosons) or by Fermi-Dirac statistics (half odd-integer spin fermions), both of which involve particles that are not distinguishable. It was in fact Dirac who first introduced the concept of spinors into QT as a means of making Schroedingers equation compatible with SR (such quanta obeyed the spin statistics of an assembly of fermions). Hence in QT all entities can be classified according to whether they obey such fermionic statistics (with its anti-symmetric wavefunction) or bosonic (symmetric) statistics. There is a whole lot of group theory which shows that half-integer spins obey the former (e.g. those whose angular momentum is 1/2h). This in turn means that 2 fermions cannot occupy the same quantum state (Paulis Exclusion Principle), an important fact since it explains the stability of the electron configurations in an atom and hence allows the richness of the periodic table. Bosons on the other hand like to be in the same quantum state and the fact that they can be stimulated to do so is exploited by the 'Light Amplification by Stimulated Emission of Radiation' (LASER). Super cooled helium-4 also obeys these statistics (in this case bosons are conserved) and thus allows the production of an Einstein-Bose condensate, the so- called fourth state of matter (what ever happened to plasma?).Whole integer spinors (bosons) faithfully represent the rotational group (they behave like the squares in a Rubic cube) while fermions exhibit a non-faithful representation, hence if you rotate an electron by one revolution it ends up being 'upside-down'.[ This property is responsible for its 'anti-social' behaviour, since it gives rise to the anti-symmetric wavefunction that characterises an assembly of fermions.] Let us take another look at the origin of this double-headed aspect of QT. A physical law needs to be invariant under rotation, hence we need to know how to express a transformation, so as to maintain the symmetry of an equation, with respect to all observers who are oriented in different directions. Wavefunctions (state vectors) exist in Hilbert space which involve complex numbers and undergo unitary transformations in order that probabilities are conserved. The special unitary group SU(2) achieves this and acts on the 2-dimensional space of spinors (so called because it represents spin states). When the XYZ axes are rotated by an angle @ (e.g. 360) the spinor rotates by only @ /2 (i.e.180). Hence we have two important representations of the rotational groups viz. O(3) acting on real vectors in Euclidean 3-D space and SU(2) acting on complex space. The later is actually simpler (simply connected) while the latter is doubly connected. Wavefunctions must transform under this (compact) unitary group in order to preserve transition amplitudes however relativity transformations involves a non-compact Lie group. For simple rotational transformations this can be achieved as already mentioned, by using a spinor representation, which under the action of the Lorentz group, undergoes a unimodular transformation and under the restriction to the rotational sub-group, it also undergoes a unitary change. The Pauli matrices generate such a special unitary group SU(2). Hence if an observers axes are rotated, the group that transforms this wavefunction must be both unimodular and unitary. This requires using spinor space not vector space i.e. as vector co-ordinates go through an Euclidean rotation, the spinors wavefunction evolves in a unitary manner. This spinor space is not however a faithful representation of the rotational group and leads to the result that 2 rotations are needed to make the spinor function return to its original state and it is this which is the cause of the Lande factor g = 2. [One rotation induces a negative sign producing an anti-symmetric wavefunction, when two particles are interchanged.] Although first rank spinors are a non-faithful representation of the rotation group (e.g. electrons) as can be shown from studying unimodular group space, higher rank spinors can produce faithful representations e.g. photons. Pauli and Schwinger showed using the CPT theorem (the invariance of physical laws under reversal of charge, parity and time), half integer spinors have operators which are anti-symmetric i.e. fermions, while integer spinors operators are bosonic. This can be achieved directly from the group theory of higher rank spinors Although not realising it at the time, Dirac was finding an equation that allowed compatibility between the unitary group of QM and the unimodular group of SR and it is this which produces g = 2, as well as invoking of anti-matter and the necessity of a field theory. [Unfortunately as Cartan has shown, such a spinor equation cannot be extended to the Riemannian techniques required by GR.] Historically, Dirac was motivated by the inadequacy of the Klein-Gordon equation. His main dissatisfaction was not that this relativistic version of QT could yield embarrassing negative values to the probability density, but rather that it did not conform to his beloved transformation theory, which he once remarked was his greatest single achievement. His contact transformations showed that QT could be expressed using a variety of bases e.g. momentum space. Just as GR was like a marble statue, which could be covered in an endless variety of covariant space-time coordinates, so could the tenets of QT be written in a countless choice of bases in Hilbert space. The interchange of two non-distinguishable particles corresponds to 360 degrees of rotation in real space, which therefore corresponds to 180 degree of rotation of a (simple) spinor wavefunction. This does not affect the probability of a wavefunction being made, since this depends on square of the amplitude of the function, but it does change the sign of the wavefunction, which means that it must be anti-symmetric. Such anti-symmetric wavefuctions implies that no two quanta can occupy the same state i.e. Paulis exclusion principle is invoked! What about the symmetric function of bosons? Well they are also unitary and unimodular, but the representation of the rotational group takes place on a space, which is the product of two spinors, which faithfully represent the rotational group. There is a great deal of group theory proving all this and it can be shown that an odd number of spinors multiplied together gives a non- faithful (fermionic) representation of the rotation group, while an even number gives a faithful representation (bosons). To recapitulate, only suitable way of representing QM, (whose wavefunction evolves/transforms according to a unitary group) such that rotations of the X,Y and Z axes, can be catered for, is to resort to spinor space. The basic spinor rotates at half the rate as that of an Euclidean vector and hence it has an angular momentum of ½h and g = 2. This also causes an interchange of particles (= 360 degrees of rotation) to give a sign change, hence the function is antisymmetric, implying that no two particles can occupy the same state (fermionic statistics). For bosons, the spinor space is made up of an even number of spinors multiplied together (e.g. 2 for photons = spin 1h) and they faithfully represent the rotation group and have symmetrical wavefunctions. Indeed an even number of fermions may itself interacts with another of its ilk, in a bosonic fashion. The more important group of transformations is the Lorentz group of special relativity (involving boosts in velocity as well as rotations). There is also a representation of this larger group in spinor space but although it is unimodular it is not unitary and the spinor has 4 components described by the transformation SU(2) * SU(2), (i.e. a pair of 2 spinors, each 2-spinor being related by parity). Moreover SR forms a non-compact group, while unitary transformations are compact and therefore SR is not compatible with basic QT. There is a theorem that states that there is no finite dimensional unitary representation of a non-compact Lie group and this overtly prevents a unitary group from representing the full Lorentz group. This leads to quantum field theory, which gets around the problem by introducing "functionals"spin functions with an infinite degree of freedom. [For example Diracs equation in which there is a four spinor composed of a pair of two spinors related by parity.] Supernumbers also allows a way around this problem e.g. superstring theory. Classically a particle has 3 degrees of freedom (X,Y & Z) while a
field has infinite degrees of freedom each position has a field value).
Quantum mechanically a particle (quanta) is represented by the
wave function which also has 3 degrees of freedom W(x,y,z). There is a
position operator and a momentum operator, which extracts eigenvalues
when acting on the wavefunction and these correspond to the classical
observables. For a (relativistic) field W(x,y,z) represents the
amplitude of probability of finding a given distribution of particles/antiparticles
of various energies and spin at that point (and takes the role of the
position operator in that of a particle Lagrangian). Whereas Quantum mechanics
allows the duality of particles and waves, quantum field theories allow
a duality between matter and energy, string theory allows a unification
of fermionic and bosonic fields. Remember that heisenberg's Uncertainty
Principle relates to the complementary uncertainties in a particles position
and momentum (~ rate of change of positioned) and when applied to fields,
it implies that the more we know about the value of a field (number of
particles at a location),.the less we know about its rate of change at
that location. A spinor field transform unimodularly and this causes the wavefunction (which has infinite degrees of freedom) to transform in the desired unitary way. [The field, being a functional behaves as if it is a function with an infinite number of variables. It therefore avoids the problem of there being no finite unitary representation of a non compact group, such as the Lorentzgroup). This ensures that observable measurements are in agreement with special relativity as well as QM. In Diracs equation, no two spinors can occupy the same energy state (i.e. fermi statistics), while for photons, the likelihood increases and their (bosonic) statistics are expressed by the commutative relationships of the field operators. As already mentioned, it was the quest to find a relativistic wave equation, which forced Dirac to introduce spinors as a necessity .[Cartan and later van derWaerden developed a complete theory of spinor calculus. The discovery of spinors came as something of a shock to the physics community, who up until then had thought that tensors were the only possible representation of the rotation group, while in actual fact if we relax the requirement of being faithful, we can allow the two-valued representation of spinors ]. This formulation also showed that particles had an inherent spin of 1/2h (electrons become inverted when rotated by 360 degrees), and the existence of anti-particles. [The prediction of anti particles, however is an abandonment of the Dirac equation as a single particle equation, since it is now required to describe both particles and antiparticles and we must therefore view it as a field equation] The only way that a spinor (which must obey a general unimodular transformation under a Lorentz transformation), can conform to the unitary transformation demanded by a quantum wave function, is if we move to what is sometimes referred to as second quantisation, in which a many particle wave function is achieved by changing the corresponding one particle wavefunction into a set of operators satisfying certain commutational relations. This results in a quantum field theory in which there is an indefinite number of particles distributed throughout space i.e; QM + SR = Quantum field theory In Dirac's equation we have moved from a function to the concept of a functional, which is a function with an infinite number of variables. As in the concept of a field, every position in space and time is given a variable value i.e we have infinite degrees of freedom. By considering functionals of spin functions and assuming that that the Hilbert space is a space of functionals, the spin can be extended so as to be in agreement not only with the group of rotations but also with the full Lorentz group! Under such a formalisation, when a Lorentz transformation is applied all the spin functions will undergo a unimodular transformation and hence behave like spinors, while spin functionals (the wave function psi) will undergo a unitary transformation, as required in order to preserve probability amplitudes. Dirac used his relativistic equation to obtain a method of describing the interaction of electrons with their electromagnetic field. This tentative way of doing electrodynamic calculations involved incorporating into the Dirac action, the Maxwell field, which although relativistic was not quantized (as was his electron field). This approach therefore considered the Dirac electron interacting with a classical Coulomb potential and although giving correct magnetic moment for the electron (Lande factor g=2), does not account for more subtle refinements such as the Lamb shift. There was already a way of expressing an electromagnetic field as an assembly of bosons, by utilizing Fock space, constructed from harmonic oscillators (together with the correct commutational relations for its creation and annihilation operators). However this was not a relativistic description of an assembly of photons. Later methods developed by Jordan, Heisenberg and Wigner etc., involving the canonical approach, did however produce a relativistic quantum field theory of electrodynamics. Alternatively, there is Feynman's path integral method which relies on a concept that was also initiated by Dirac. He elaborated it so as to incorporate interactions with quanta of electromagnetic fields (photons) as well as virtual particles and antiparticles of its own electron field. Quantum electrodynamics [QED], does however suffer from embarrassing infinities which occur when probing interactions at higher energies and these have to be removed by a process known as renormalization (a technique that does not however work for quantum gravity). Electrons however, are not the only source of electric charge and a more complete description would allow for quarks! Such Quantum Flavour and Colour Dynamics [QCD &QFD] that are incorporated in the Standard Model, rely on gauge theories in which there is spontaneous symmetry breaking that is crucial in explaining the mass of the gauge particles that carry the interaction (and indeed possibly explain the fermionic masses in Grand Unified Theories) More generally, in order to quantize a classical field, we first express it in a Lagrangian form and find its corresponding (position and momentum) canonical conjugate variables -- the first of which I will denote as psi(x,t). [The required classical equation for the field is then usually retrieved by using Hamilton's equations]. Then we quantize the field by subjecting the canonical field variables to the correct Heisenberg's commutational relationships. The equation of motion for any quantum variable 'F' can then be obtained by replacing the Poisson bracket by the corresponding Heisenberg commutator bracket (which can be found since F and the Hamiltonian are given in terms of the known canonical variables). The quantum field quantity psi(x,t) is then regarded as an Hermitian operator (rather than a real numerical function) whose Fourier expansion can be written in terms of its creation and anihilation operators and in a relativistic field, we have to cater for both matter and ant-matter. These two operators must also satisfy commutational relationships, depending on whether the field represents an assembly of bosons or fermions.The quantity psi(x,t) plays a role in field theory analagous to that played by x, the position vector in particle (quantum) mechanics and the two canonical variables obey the same commutational rules of Heisenberg. The process of regarding ps(x,t)i as an operator rather than a number, is therefore part and parcel of the process of second quantization (i.e. the quantization of a field.) This field quantization has an obvious interpretation as a many particle theory, where the square of the amplitude is proportional to the number of particles present. An alternatively method of providing a quantum field theory, is to apply the path integral method and use Feynman propagators ( which are the inverse of the operator appearing in the quadratic part of the Lagrangian), to determine the contribution of each of the scattering probabilities of the particles interacting with their field. Supersymmetry involves combining fermions (the basic constituents of matter) and bosons (responsible for the fundamental forces) on an equal footing in the same quantum field theory. As well as unifing fermions with bosons, it also unifies spacetime symmetries with internal symmetries, and (in the case of local supersymmetry), gravity with matter. The non compact groups of SR do not have a unitary representation in commuting numbers (e.g complex or real space) but there is a possibility in superspace. Bosons arise naturally as a field which provides the interaction between fermion particles, which usually manifest as a force (electromagnetism, strong nuclear interaction and gravity). Different fermions (e.g. electrons, neutrons) can be represented as an internal state, in which these components (fermions) are symmetric under certain local gauge transformations. [This state behaves like spinors and the gauges like SU(2)]. However the boson field associated with gravity (gravitons of spin 2h), must belong to the none compact group of SR, since GR which describes gravity, is a localised (hence generalised) form of SR. As mentioned above, we therefore need to resort to anti-commuting numbers in order to achieve this master supersymmetry and obtain unification. Fermions and bosons are therefore united in the same Quantum field theory and can be interchanged into each others state by symmetrical internal transformations e.g. electrons-sleptons or gravitons- gravitinos. Supersymmetry is an extension of ordinary Poincare space-time symmetry which is obtained by adjoining N spinorial generators whose anticommutator yields a translation generator (hence producing gravity when localized). This symmetry can be realized on ordinary fields (functions of space-time) by transformations that mix bosons and fermions but a more compact alternative to this component field approach is given by the superspace- superfield approach. Here superspace is an extension of ordinary space-time so as to include extra anticommuting coordinates in the form of N two component Weyl spinors $. Superfields Psi(£$) are then functions defined over this space. The transformations mixing bosons and fermions are then constant translations of the $ coordinate and also related rotations of the $ into the space-time coordinate £ In (closed) superstring theory, upon quantization, the canonical conjugate operators decompose into a Fourier series, which contain left and right handed harmonic oscillators which do not interact (that is as the string propagates, it has distinct right and left moving oscillator modes). Quanta are therefore represented by the dynamics of strings and different vibrational modes in different bosonic/ fermionic coordinates represent different quanta. The left moving modes are purely bosonic and exist in a 26 dimensional space, which has been compactified to 10 dimensions. The right moving modes only live in this 10 dimensional space and contain the supersymmetric part of the theory, this requires the introduction of superpartners that differ by a spin half (e.g. electron-slepton, photon-photino, etc) as well as the graviton/gravitino. The compactified 16 dimensional string lives on the root lattice space (e.g. 16D Tori), of an E8*E8 isospin (internal) symmetry, which is more than large enough to contain the required spectum of particles. When the left moving half and the right moving half are put together they produce the heterotic string (meaning "hybrid vigour"). Compactification of the extra six dimensions on a Ricci flat (e.g Calabi-Yau) manifold, then reduces the 10 dimensional superstring into our familiar 4 dimensional space-time, breaking the E8*E8 symmetry to produce the electroweak and strong interactions that are represented by U(1), SU(2) and SU(3) gauge symmetries. [ Under compactification that produces the Calabi-Yau space, one of the E8 groups is broken down to its E6 subgroup and E6 is a good candidate for a Grand Unified Theory, as it contains the subgroups of the Standard Model. The other E8 sector would describe (shadow) matter that could (only) interact gravitationally with the 'E6 matter'] Hence although the reason that fermions and bosons exist in subtle, deep and complex, if these symmetries were not enforced upon nature, we would not have the richness that we observe in the universe. This is nowhere more evident than in the creation of our own planet Earth and in the very laws that are responsible for its existence. Regarding superunification it is not possible to satisfactorily incorporate gravity (which is governed by the non-compact Poincare group) in what is known as a Unitary representation (that dictates the other 3 quantum interactions), unless one resorts to supernumbers, which combines both fermions and bosons via supersymmetry. This concept was originally invoked in the early study of string theory but although this has received a recent surge in popularity, supersymmetry itself does not require a string formulation. I'll let Dirac have the last (controversial) word "Physical laws should have mathematical beauty" (epitaph)
* * * * * * * * * * * * * * * ADDENDUM Below is a summary of the 5 versions of String Theory Preamble String theory can trace its origins from a study of S-matrix theory (S
for scattering), which was initiated by Heisenberg and championed by John
Wheeler and later Geoffrey Chew. This approach allows the description
of quantum interactions in terms of inputs and outputs, without a detailed
knowledge of what was actually occurring 'In the black box' but merely
relying upon certain constraints of relativity symmetry and quantum theory
etc. Gell-Mann had noticed regularities, so called T dualities (symmetries)
present in these S-matrix approaches and this was later consolidated by
Veneziano when he discovered that it obeyed the Euler beta function. Still
at this stage no one knew what was the mechanism behind these scattering
diagrams and understanding such a mechanism was crucial to understanding
the strong force which dominated such collisions and the ever increasing
plethora of newly discovered particles (this was before the advent of
QCD and quarks). It was however later discovered by Nambu, Susskind et
al, that the mechanism could be explained by resorting to a string rather
than the ubiquitous particle description (of the Lagrangian). This Bosonic
strings was found to require 26 dimensions in order for the theory to
obey the requirements of special relativity. However such a string theory
was employed to describe the hadrons which were predominantly fermions.
In 1971 Ramond found that by employing spinors, he could derive a string
theory for fermions in 2 dimensions and in doing so, discovered an early
form of supersymmetry [SUSY]. [Supersymmetry was originally used in an
attempt to produce a Master group, that would allow both internal (isospin)
groups and noncompact space-time groups to be combined in a non-trivial
fashion Mandula and Coleman had proved that the Poincare group of translations,
rotations and boosts is the maximal symmetries applicable to space-time
but this did not take into account the possibility of supermanifolds].
however Ramonds superstring theory had only one dimension of space and
one of time but in 1973 Wess and Zumino developed a 4 dimensional supersymmetric
(particle) theory. Then in 1976 this theory was extended so as to include
gravity - - -producing supergravity! [Basically since their spins are
different, bosons and fermions transform differently when they rotate
in space. Supersymmetry transformations must involve space and time in
order to compensate for this distinction. In actual fact a local supersymmetric
transformation (as decreed by SR), requires a localised transformation
in space-time, that is inherent in Einsteins General Relativity
theory of gravity.] By 1974 string theory entered a dry period which, saw the success of quantum chromodynamics in explaining new strong interaction phenomena. What followed was the rise of the standard model and later supergravity, which were considered more compelling theories and retained the more familiar particle viewpoint. Interest in superstring theory underwent a revival when in 1984 Schwarz and Green showed that certain unwanted terms that led to anomalies (these are symmetry violations such as the conservation of energy), miraculously canceled. These symmetries must be exact otherwise there is no way to eliminate the unwanted polarization of the gauge boson. For example, the photon only allows 2 transverse polarizations and the unwanted longitudinal direction is forbidden by internal symmetries - - - simmilar internal symmetries restrict the W, Z and gluon gauge bosons. [Group theory of Wigner et al, shows that massles particles only have 2 polarizations, while massive gauge bosons have 3. This is because massless particles travel at the speed of light and are therefore never at rest. Consequently they single out a particular direction of motion and hence distinguish the perpendicular motion from the remaining polarization along the direction of travel. It turns out that for massless gauge bosons, physical polarization oscillate only in the 2 perpendicular directions.] Hence Schwarz and Green showed that the sum of the quantum mechanical contributions to all these possible symmetry breaking anomalies in 10D superstring theory is zero Also by radically reducing the size of the string, they could make a leap forward in unification by identifying the spin 2 boson as the graviton. Superstring theory also received a boost from the work of Witten, who produced several new results, such as the identification of Calabi-Yau spaces as candidates for the extra curled up dimensions and the dependence of the number of generations of particle to the Euler number. [The precise size and shape of the extra dimensions has a profound effect on string vibrational patterns and hence on particle properties. Calabi-Yau spaces produced the left handed chirality that is needed to explain weak interactions which violate parity by distinguishing between left and right handed interactions.Furthermore, rolling up the extra dimensions into a Calabi-Yau manifold preserves supersymmetry by just the right amount; too little prevents supergravity, too much and we wouldn't get left handed particles that had different interactions from the right handed ones]] One of the appeals of superstring theory is that it overcomes the conflict of QT with GR since it smears out space time in a way that causes the microcosm to come to an end near the Plank scale and so obviates the embarrassing infinities that had plagued other methods. Also, all the fundamental quanta ( both particles and their messengers) can be explained as various vibrations of the same string theory. Also if a huge collection of strings all vibrate in just the right coordinated way throughout all of space, they can provide a uniform background that is similar to the Higgs ocean and these effects should be calculable as opposed to imputed as in previous particle versions [Note that only the zero mode of a string can be associated with the known particles, since the first excitation of a string has a mass equal to that of the Plank mass (and multiples thereof for higher harmonic vibrations), which is far too large to be created in particle accelerators. These zero mode vibrations would then have to pick up masses by a Higgs mechanism, so as to produce the observed spectrum of fermions etc.]. String vibrations that initially yielded zero mass, would then acquire tiny nonzero masses through the drag force they experience as they move and vibrate through the string version of the Higgs ocean. In actual fact supersymmetry helps solve the Hierarchy problem since it allows a light Higgs particle when there are large quantum contributions to its mass from virtual particles. (These contributions would otherwise require a 'fudge factor' which needs to be continually finely tuned at each energy scale). with SUSY the contributions from both partner and superpartner cancell out (fermion contribution is negative while bosons is positive). Supersymmetry also ensures that the various strengths of the interactions converge to a common value at a high enough energy, a feature that is necessary for unification but is inexact in other GUTs. However it transpired that there could be 5 variations of 'string theory' and that each of them was strongly effected by the topology of the manifold. Although the orbifolds/Calabi-Yau was the most likely, this allowed many thousands of possibilities without much guiding principles as to how to proceed. The prompted the next revolution in string theory M-THEORY, in which all 5 versions of string theory are seen as different perspectives of a higher 11 dimensional Brane theory. [Note that string theory is pertubative since it requires the breaking apart interaction and reforming of strings, whereas M-theory is non pertubative]. Witten was able to show that the approximate (pertubative) string theory equations that are written in 10 dimensions, were actually 5 approximate descriptions of a single underlying 11 dimensional theory. This whole enterprise has therefore lead to a revolution in geometry, comparable to the that of Gauss and Riemann's transformation of the Euclidean geometry that had dominated for 2 Millennia. In addition to considering microscopic p-branes, more recent developments have lead to theories involving D-Branes (which could contain strings wholly within them or alternatively just restrain their ends) and macroscopic Brane Worlds (with or without strings).
Technicalities Up to now, String Theory is only well understood at the perturbative
level. The field theory diagrams are replaced Lorentz invariance on the target space or conformal invariance on the
worldsheet fixes the number of spacetime dimensions (twentysix for bosonic
strings and ten for superstrings). As our low energy world is four dimensional,
String Theory incorporates the Kaluza-Klein idea in a natural way. The
effect of a non-zero finite radius R for the compactified dimension is
just a tower of Kaluza- Klein states with masses n/R. But in String Theory,
the string can wind m times around the compact dimension. This process
gives a contribution to the momentum of the string proportional to the
compact radius, mR/a'. These quantum states become light for R --- = 0.
The dimensional reduction of a String Theory in D dimensions is another
String Theory in D dimensions. This is T duality which is discussed below.
*!* [Under spacetime compactifications, the
type IIA and For the bosonic string, the lowest vibrating mode correponds to a tachyon (a 'faster than light' particle with an imaginary mass). It indicates that we are performing perturbation theory arround an unestable minimum. Supersymmetry gives a very economical solution to this problem. In a supersymmetric theory the Hamiltonian operator is positive semi-definite and the ground state has always zero energy. It is also very appealing from the point of view of the cosmological constant problem. Furthermore, supersymmetry also introduces fermionic degrees of freedom in the physical spectrum. [Superspace is obtained by adding {four} spinor degrees of freedom to the spacetime coordinates xµ.] As already mentioned, another important consequence of string theory, is the prediction on the number of dimensions of the target space where the perturbative string propagates. Lorentz invariance on the target space or conformal invariance on the worldsheet fixes the number of spacetime dimensions (twenty six for bosonic strings and ten for superstrings). As our low energy world is four dimensional, String Theory incorporates the Kaluza-Klein idea in a natural way. But again the one-dimensional nature of the string gives a quite dierent behavior of String Theory with respect to field theory. The dimensional reduction of a field theory in D spacetime dimensions is another field theory in D-1 dimensions. The effect of a non-zero finite radius R for the compactified dimension is just a tower of Kaluza- Klein states with masses n/R. But in String Theory, the string can wind m times around the compact dimension. This process gives a contribution to the momentum of the string proportional to the compact radius, mR/a'. These quantum states become light for R approaching. zero. The dimensional reduction of a String Theory in D dimensions is another String Theory in D dimensions. This is T duality*!* The final important consequence comes from the cancellation of spacetime anomalies (gauge, gravitational and mixed anomalies). It gives only the following five anomaly-free superstring theories in ten spacetime dimensions. Type I SO(32): The SO(32) and E8 × E8 heterotic strings. We see that the Heterotic theories don't contain D-branes. They do however
contain a fivebrane soliton which is not a It is worthwhile to note that the E8 x E8 Heterotic string has historically
been considered to be the most promising string "Gravity exists, so if there is any truth to supersymmetry then any realistic supersymmetry theory must eventually be enlarged to a supersymmetric theory of matter and gravitation, known as supergravity. Supersymmetry without supergravity is not an option, though it may be a good approximation at energies below the Planck Scale. Supergravity is itself only an effective nonrenormalizable theory which breaks down at the Planck energies. So if there is any truth to supersymmetry then any realistic theory must eventually be enlarged to superstrings which are ultraviolet finite. Supersymmetry without superstrings is not an option. Superstring theory is itself only a perturbative theory which breaks down at strong coupling. So if there is any truth to supersymmetry then any realistic theory must eventually be enlarged to the non-perturbative M-theory, a theory involving higher dimensional extended objects: the super p-branes. Supersymmetry without M-theory is not an option." . . . . . . . S. Weinberg The following results helped enthuse the superstring revolution 1) Green and Schwarz discovered that the gravitational and Yang- Mills
anomalies of the ten-dimensional superstrings all cancel provided the
gauge group is either SO(32) or E8 E8;
Schematically we can summarise heterotic string theory as follows The simplest solution is of course 26D flat spacetime with constant values
of all the fields. For this case we have a 2D free (1) . . . . . . .Xm(u,v) = Xm,r(u-v) + Xm,l(u+v)
Since this is a free theory, quantization assigns canonical commutation
relations to the Fourier coefficients like the (2) . . . . . . . . . . M^2 = N,r + N,l - 2
The instability due to the tachyon can be easily cured by supersymmetrizing
the theory. In that case the tachyon state is In order to construct string models in less than 10D as well as to understand the heterotic string construction, we need to consider the simplest compactifications which correspond to the extra dimensions being circles and their higher dimensional generalization. Let us first see the case of a circle S1 This means that the 10D space is represented by flat 9D spacetime times a circle S1. We know that a circle is just the real line identifying all the numbers differing by 2PiR, where R is the radius of the circle. So the only difference with the flat space discussed above are the boundary conditions. The solution of the wave equations are now as in (2). But now Pr = m/2R - nR and Pl = m/2R +nR, . m is an integer reflecting the fact that the momentum in the compact direction has to be quantized in order to get single-valued wave function. The integer n however refers to the fact that the string can wind around several times in the compact dimension and is thus named the `winding number'. The mass formula is then: (3) . . . . . . . . . M^2 = Nr + Nl - 2 + m^24R^2 +n^2R^2 . . . . . . . . . . . Nr - Nl = mn
(4) , , , , , , , R<-> 1/2R . . . . .m <->n
The five superstring theories appear to be very different when viewed
in terms of their descriptions in weakly coupled The next duality that we will consider is called S-duality. Simply put,
this duality relates the strong coupling limit of one theory
T Duality *!* A symmetry whereby a string theory compactified on a small circle and
a string theory compactified on a large circle give rise to identical
physics. The basic idea of T duality (T for turns) can be illustrated
by considering a compact dimension consisting of a circle of radius R.
In this case there are two kinds of excitations to consider. The first,
which is not special to string theory, are Kaluza--Klein momentum excitations
on the circle, which contribute (n/R)2 to the energy squared, where n
is an integer. Second there are winding-mode excitations, due to a closed
string winding m times around the circular dimension, these are special
to string . . . . T =Lst^2/2Pi denotes the string tension (energy per unit length), the contribution
to the energy squared is . . . . R <-> Lst ^2/R This is part of an exact map between a T-dual pair A and B. Hence T-duality
states that nothing changes if we interchange the roles of the Kaluza-Klein
and winding particles providing we also exchange the radius of the circle
R by its inverse 1/R. S Duality A symmetry whereby a string theory with a large coupling constant and a string theory with a small coupling constant give rise to identical physics. Suppose now that a pair of theories A and B are S-dual. This means that if f denotes any physical observable and £ denotes the coupling constant, then; . . . fA(£) = fB(1/£) (The expansion parameter a introduced earlier corresponds to £)
This duality, whose recognition was the first step in the current revolution,
generalizes the electric-magnetic symmetry of Maxwell theory. Since the
Dirac quantization condition implies that the basic unit of magnetic charge
is inversely proportional to the unit of electric charge, their interchange
amounts to an inversion of the charge (which is the coupling constant).
So S-duality (S for strong -weak), allows the roles of the Noether (electric)
charge and the topological (Bianchi) charge (magnetic monopole) to be
reversed. Therefore if we have an electric charge of quantised value me
and a magnetic charge of value n/e (to be consistent with Dirac's
theory), then S-Duality states that if we change the integers m
and n, we must also exchange e with 1/e and go from
a regime of weak coupling to a regime of strong coupling! [It can also
be shown that in M-Theory, the electric charge of the fundamental string
and the magnetic charge of the solitonic fivebrane obey a Dirac quantization
rule. The fivebrane is the `magnetic' dual of the `electric' membrane
in D=11, in agreement with the general formula that the dual of a p-brane
is a ~p-brane with ~p = D-p-4 .For each (p+1)-form in the Lagrangian there
is an associated electric-type p-brane solution and a magnetic-type (6
-p) brane solution, carrying Qp and Q(6-p) charges respectively. The understanding of how the IIA and HE theories behave at strong coupling, which is by now well-established, came as quite a surprise. In each of these cases there is an 11th dimension that becomes large at strong coupling, the scaling law being . . . . . L11 = £^2/3 In the IIA case the 11th dimension is a circle, whereas in the HE case
it is a line interval (so that the eleven-dimensional space-time has two
ten-dimensional boundaries). The strong coupling limit of either of these
theories gives an 11-dimensional T-duality transforms electrically charged winding states into electrically charged Kaluza-Klein states, but S-duality transforms elementary electrically charged string states into solitonic monopole and dyon states (dyons carry both electric and magnetic charge). T and S duality can be combined under a so called U-Dualty Mirror Symmetry This applies when trhe 6 dimensions are rolled up into 2 different Calabi-Yau manifolds, yet the resulting 4D space-time long distance theory can be the same The mirror manifold of a given Calabi-Yau space could be entirely different in shape size twisting or genus D-Branes What about the nonperturbative effects in String Theory?. Does String
Theory incorporate nonperturbative excitations (string solitons)?. Are
there also strong-weak coupling duality transformations in String Theory?.
Before the role of D-branes in String Theory were appreciated, the answers
to these three questions were not clear. For instance, it was known, by
the study of large orders of string perturbation theory, that the nonperturbative
effects in string theory had to be stronger than in field Another source of insight into non-perturbative properties of superstring
theory has arisen from the study of a special class of p-branes called
Dirichlet p-branes (or D-branes for short). The name derives from the
boundary conditions assigned to the ends of open strings. The usual open
strings of the type I theory satisfy a condition (Neumann boundary condition)
that ensures that no momentum flows on or of the end of a string. However,
T duality implies the existence of dual open strings with specified positions
(Dirichlet boundary conditions) in the dimensions that are T-transformed.
More generally, in type II theories, one can consider open strings with
specified positions for the endpoints in some of the dimensions, which
implies that they are forced to end on a preferred surface. At first sight
this appears to break the relativistic invariance of the theory, which
is paradoxical. The resolution of the paradox is that strings end on a
p-dimensional dynamical object -- a D-brane. D-branes had been studied
for a number of years, but their significance was explained only recently. S ~~[(XuXu -2$u$u) - {£a£a where X =Boson field, ('u' and 'a' are dimensions of 10 and 16 respectively) $ =Fermion field (2 component world sheet spinors) propagating along string and transforming as a vector representation of the Lorentz group The heterotic string has the structure of an E8*E8 {or SO(32) } gauge symmetry and D=10 spacetime supersymmetry. (Simple bosonic strings were found to need 26D but superstrings require 10D) The physical space is a tensor product of the Right moving oscillator modes (fermionic) that include supersymmetric degrees of freedom (10D). [these involve the right moving modes $u and the right moving part of Xu] and those left moving modes (bosonic 10D) that are responsible for Gauge symmetries [these are the left moving part of Xu and the £a.] Red = 10 dimensional field in which the Right moving oscillator modes are fermionic (both Xuand $u have these) and the Left moving modes (Xu field only) are bosonic. (The left moving 10D bosonic string combines with the right moving bosonic string to complete a bosonic closed string).Because of the Right moving modes there is consequently a supersymmetry between the bosonic Xu and the fermionic $u ..as in typeII string models. . viz &Xu=ie$u_ ,,,,,, ,,, and &$u_=e&_Xu. where e is the supersymmetry generator Bold =16 dimensional fields which contain only Left moving modes £a, therefore it only has gauge symmetry, (probably an E8*E8 isospin symmetry of a compactified 16 Tori root lattice space) £a = originally 26D fermion coordinate field that can be replaced by 16D bosonic coordinates that transform with isospin gauge symmetry ie. containing internal quantum numbers. Together with the 10 dimensional left moving bosonic field of Xu, these produce the 26 free bosons necessary in the Veneziano string model. The 10D supersymmetric space is itself compactified into a Calabi-Yau manifold and 4D spacetime, which reduces the E8*E8 gauge symmetry into the subgroups of U(1)*SU(2)*SU(3) that occurs in the standard model (hopefully complete with graviton field and anomally cancellations).
The string action, is a field theory in 2 dimensions and Xu(#1,#2) where u =1 up to D-1 dimensions are coordinates for a string that is propagating in D space-time dimensions. A more generalized 2 dimensional field theory that would correspond physically to internal degrees of freedom that are free to propagate along the string, such as the free fermionic field $a(#1,#2), which could transform in the vector representation of the Lorentz group SL(D-1, 1). It may seem counterintuitive that an anticommuting field should transform as a vector -- a bosonic representation -- of SL(D-1, 1). This choice means that $a(u=10) maps bosons to bosons and fermions to fermions in the space-time sense (remember that there is however a supersymmetry invoked by this introduction). However since this is a 2 dimensional field theory, not a field theory in space-time, and $a(u=10) transforms as a spinor under transformations of the 2 dimensional world sheet, in perfect agreement with the usual spin statistics. The Lorentz group SL(D-1, 1) is merely an internal symmetry from the world sheet point of view and hence the spin and statistics theorem says nothing about whether anticommuting fields should transform as vectors or spinors under an internal symmetry. In this manner we associate Lorentz quantum numbers to $. In addition we can attribute other internal quantum numbers (as is necessary in heterotic theory) by adding a 32D real Majorana 2 component spinor, which transforms internally under a SO(32) gauge symmetry. This then insures the canceling of any Virasoro-ghost anomaly since 2 Majorana spinors are equivalent to one boson anomaly. Alternately and more interestingly, while letting the $ fermionic field transform internally as a Lorentz quantum number, we can account for the necessary 26D bosonic requirement by introducing a 16D bosonic left moving mode £, whose internal degres of freedom are subject to the E8*E8 symmetry caused by compactification onto a 16 tori. The isotopic symmetry thus produced can then account for the required spectrum of particles. Hence the 2D world sheet string propagates in terms of parameters #1 and #2 as it passes through a multidimensional space that includes a 10D space-time, that has bosonic and fermionic field components (i.e. it forms a supermanifold), which together eliminates tachyons. The fermionic field is is made of 2 component world sheet spinors - each being a Lorentz singlet (the world sheet parameters #1, #2 have a 2 spinor associated with them, which can be mapped to a 10D fermion field that is space-time covariant). There is also a 16D bosonic field, whose oscillator mode is needed in order to eliminate ghosts (bosonic strings theories requiring a total of 26D), and this can also be expressed as a 32 component spinor field (2 Majorana spinors = 1 boson Virasoro-ghost anomaly) i.e. an SO(32) gauge symmetry, or as is more favorable we could use the E8*E8 symmetry of a compactified 16 Tori. The gauge symmetry therefore connects isospin particle ie. fermions that carry internal quantum numbers. The left moving modes are bosonic (producing gauge transformations) while the the right moving modes are fermionic (producing supersymmetry) The left/right asymmetry that is known to occur in the weak interaction (and therefore the standard model), unfortunately leads to anomalies in string theory i.e. there is an unwelcome breaking of the symmetry when the string is quantized. Only two symmetries are known to survive such necessary quantization viz. S)(32) and E8*E8. However at present we usually can only apply perturbation theory and his allows too many solutions. An exact method may isolate just one unique string theory -- the one that accurately describes the vacuum, without yielding an embarrassingly large value for the cosmological constant! * * * * * * * * * * * * More recent developments; Braneworlds The Higgs mechanism makes it look as though the weak force symmetry is preserved at short distances (high energies) but is broken at long distances (low energies).The weak Higgs field has 2 components which are both zero when the weak symmetry is preserved, however when one of these takes on a nonzero value it breaks the weak force symmetry that interchanges the 2 Higgs fields. The symmetry is broken spontaneously because all that breaks it is the vacuum -- the actual state of the system, the non zero field in this case. However the physical laws which are unchanged, still preserve the symmetry even though the physical system does not. The symmetry transformations that act on the weak gauge bosons, also act on the quark and lepton flavours and it turns out that these transformations wont leave things the same unless they are masslees. Now because the weak force symmetry is essential at high (GUT) energies, not only is spontaneous symmetry breaking required for the gauge bosons, it's necessary for these quarks and leptons to acquire mass as well. The Higgs mechanism is the only way for all the massive fundamental particles of the standard model to acquire their masses. At high energies the internal symmetries associated with the weak force, still filters out the problematic polarization of the weak gauge bosons that would cause interactions at too high a rate. However at low energies, where the mass is essential to reproducing the measured short range interactions of the weak force, the weak force symmetry is broken. In this mechanism, there are original 3 weak gauge bosons plus a Goldstone boson and after symmetry breaking we have the massive W+- and Z bosons plus the photon, which is able to travel in a massless mode through the Higgs vacuum unaffected, since it has no weak (flavour) charge. The problem with GUT is that although the Higgs particle has to be relatively light for weak symmetry breaking (~ 250GeV as born out by experiment,) it is partnered another particle (X), that interacts with it through the strong force and which has to be extremely heavy [This in order to explain the stability of the proton, which would otherwise decay due to the X particle allowing a quark to change into a lepton]. In other words we are left with the problem that 2 particles that are related by GUT symmetry have to have enormously different masses (the weak and strong forces have to be interchangeable at high energies). This hierarchy problem is made worse by the fact that QT requires the value of the Higgs particle to be determined by contributions of virtual particles which are of the order of the energy scale of GUT (according to the anarchic principle in which any interactions that is not forbidden by symmetry will occur). However these QT contributions (some positive some negative), which must be added to the classical value of the Higgs boson, are under GUT energy values 13 orders of magnitude greater than the weak Higgs value! The situation is even worse when we consider gravity since QT corrections now occur at the Planck scale (10^19Gev) [Newton's law states that strength is inversely proportional to the square of of the energy/mass and because gravity is so weak the Planck scale is large Another way of phrasing the hierarchy problem is to ask why gravity is so feeble]. Such a trickle down effect of QT contributions should therefore make a large quantum mass determine the ultimate mass of other particles, so that all end up rich in mass. Supersymmetry (SUSY) gives an answer to this by allowing the positive QT contributions of the bosons to be canceled by the negative contributions of the fermions. It achieves this by first pairing all the fermions and bosons with superpartners, the Higgs field then gets contributions from both particles and supersymmetric particles and because the interactions with the two are different their contribution to the Higgs particle's mass, cancel each other out. Now in order to account for the lack of observed superpartners, it is necessary to invoke a SUSY breaking, which imbues mass to the superpartner, making it to large to to be stable or created in particle accelerators. However once supersymmetry is broken, flavour changing interactions are allowed which are not observed in nature or at least are a lot more rare than predicted. These are processes that change quarks or leptons into those of another generation (that is ones that are heavier or lighter but with the same charge). Although an electron and a slectron can interact via the weak force as can a muon and smuon, an electron would never interact directly with a smuon. If an electron were paired with a smuon or a muon with a slectron, this would allow a muon to decay into an electron and a photon, something which is never observed. However with SUSY breaking the now massive bosonic superparteners no longer have the strong sense of identity of their partner fermions, and this allows the (massive) bosonic superpartners to get all mixed up, so that not only a smuon but also a slectron would be paired with a muon for example. However this pairing of a slectron and a muon would yield all sorts of interactions that are not observed. So although SUSY can overcome the hierarchy problem it does lead to a flavour problem.The question is how do we break the SUSY, but prevent the flavour problem from occurring. One possibility for resolving the called flavour problem is to resort to large dimensional branes (without SUSY), which may also allow a remedy to the Hierarchy problem. An alternative approach to SUSY in addressing these and other problems, is Braneworld theory.Branes originated from string theory (1989), in which D-branes (Dirichlet boundary conditions), were used as ends for strings that move in the bulk space. At around the same time p-branes were discovered as solutions to Einstein's field equation in higher dimensions. These extended infinitely far in some spatial dimensions but in the remaining dimensions they act as black holes. In some geometries they are found to give rise to new types of particles that are not accounted for in D-brane string theory. These p-branes are independent objects that can wrap around a very small curled up region of space time, an act like particles. In 1995 it was shown that D-branes and p-branes were actually the same thing and that at energies where string theory makes the same predictions as General Relativity, D-branes morph into p-branes. The way in which this equivalence is best expressed is via the important notion of duality(cf. *!* above.) An important aspect of duality was revealed by Witten in 1995 when he demonstrated that a low energy version of 10D superstring theory with strong coupling was equivalent 11D supergravity with weak coupling (which could therefore be dealt with by perturbation techniques.Dualities between all the contending string theories were established bringing about a second revolution in string theory, namely M-Theory. In order to reconcile the difference in dimensions it was realised that the strings were actually membranes that extend in dimensions that were previously not recognized due to their compact size and these have been identified as the p-branes. [Eleven dimensional supergravity although not containing strings was already know to contain 2D membrane solutions] The key was therefore the realisation that rolled up dimensions are invisible at long distances or low energies, making 11D supergravity with one dimension curled up equivalent to 10D string theory [In 11D supergravity you need to know the momentum in 10D whereas in 11D superstring theory you need to specify the momentum in 9D and also the value of the charge, i.e. 10 numbers have to be specified in each case to make the particles correspond in the 2 theories. [Perhaps this is an indication that these extra dimensions should be regarded as auxilliary variables just as is charge} Ordinary uncharged strings do not pair with objects in 11D and the partners of objects in 11D theory turn out to be branes viz. charged pointlike branes called Do-branes. The 2 theories are dual because for every Do-brane of a given charge in 10D superstring theory, there is a corresponding particle with a particular 11D momentum and vice versa] It was soon realised that it was possible to formulate theories of higher dimensional branes within the context of M-theory. Mmathematicians also began to consider the possibility of a higher dimensional universe in which the particles and forces reside in a lower dmensional brane - - a Braneworld! The first example of this was the HW braneworld (Horava-witten), in which 2 branes bounding the 11th dimension, were shown to be equivalent to the heterotic string, with strong coupling (this is yet another example of duality). the new feature of braneworlds is that it allows particles/forces to exist on seperate branes and only able to communicate weakly via bulk particles such as the graviton. Although strings representing particles and forces can be trapped to branes there is no requirement to resort to them but braneworld theory does assist SUSY in solving the hierarchy problem. [The difficulty with this approach is that the SUSY needs to be broken if is to explain why we observe particles but not the (massive) superparners.] This problem results from the anarchic principle, and is due to the (virtual) quantum contributions (especially at the GUT or Planck scale), to the Higgs particles mass (which after all is responsible for imbuing mass to all other particles). An alternative way out of this problem, is to assume that sequestrating particles on a separate brane can prevent these unwanted interactions that until now could only be restricted by symmetries. Basically by sequestrating the unwanted particles on a seperate brane the anarchic principle can be restricted. The graviton (or maybe the gauginos) are able to travel through the bulk and are responsible for communicating the SUSY breaking but since this breaking happens sufficiently far away it will have very little effect. This graviton induced communication of SUSY breaking is known as anomaly mediation. In this way the interactions of the Standard Model remain the same as in a theory with unbroken symmetry. So just as in a theory with exact SUSY unwanted flavour changing do not occur. [A refinement of this concept is that other particles responsible for flavour symmetry breaking are sequestrated on other branes, the breaking being communicated to Standard Model particles only via particles in the bulk Different flavours of quarks and leptons would be different because they interact with a different brane at a different distance (the further the distance, the smaller the mass induced), an effect termed shining] These bulk particles although originating/travelling in higher dimensions should still leave some sort of trace in our 4D world and are refered to as Kaluza-Klein KK particles (a term which is also used in the T dualities in M-theory, since in both cases, their energies a determned by the size of the extra dimensions - - branes may be curled up in M-theory ). If the extra dimensions are bigger the KK particles are lighter; if the world is higher dimensional but with no branes then all familiar particles will have KK partners. They would have exactly the same charge but carry momentum in the extra dimensions. The next stage in development was the realization that extra dimensions rather than SUSY could explain the Standard model. This ADD idea postulates that larger dimensions might explain the (apparent) weakness of gravity and explain why the Planck scale and weak scale are so different - - thus solving the hierarchy problem [The hierarchy problem can be expressed in another way; why is the Planck scale so large when the weak scale is so small or equivalently why is the strength of gravity so weak] Essentiall the ADD model claims that the fundamental mass scale that determines gravity's strength is not the Planck scale mass but one much smaller (gravity's strength is inversely proportional to the Planck scale) In these and subsequent developements , the larger the extra dimension, the weaker the gravitational force in the effective lower dimensional universe would be. The initial spreading out in the extra dimensions would reduce the density of the force lines in lower dimensional space, so the strength of gravity experienced would be weaker. If a Tev were the typical energy of gravity, there would no longer be a hierarchy of masses in particle physics. the Planck scale mass that we measure in 4D is large only because gravity has been diluted in 'large' extra dimensions. If these ideas are true then it means that it would be possible to create black holes and KK particles at energies close to a Tev and also we may be able to detect a deviation of Newton's inverse square law at small distances. Further developments involved warped passages connecting these branes - - the RS1 theory of Randall and Sundrum [a warped space is one in which the cross sectional dimensions have the same geometry (e.g. flat), while the extra dimensions have a variable curvature]. This comes about when we consider branes which themselves and the bulk space, have energy and can therefore induce a curvature as we move away into the extra dimension. By solving Einstein's field equation for a particular braneworld we obtain a gravitational probability function in which the strength deteriorates the further we move away from one of the branes gravity is seen to be confined on one brane (the Gravitybrane which carries positive energy ) and weakens as it heads towards the weakbrane (which carries negative energy) upon which all the standard particles reside. The warped space causes gravity to decrease at an exponential rate and in this way we do not require a large separation between the 2 branes and no contrived large number in order to explain the hierarchy. The huge ratio of the Higgs masses is expected if gravity (Planck scale), is confined near the gravitybrane, while the weak interaction is confined on the weakbrane some 16 units away (giving a reduction by 10^16) In order to accommodate the apparent unification of the interactions that is indicated by the convergence of the strengths at higher energies, work has revolved around the application of strings which are allowed to move through the bulk These strings represent the various standard model particles and their gauge bosons. Unlike the graviton which must arise from a closed string, gauge bosons and fermions will be either stuck on a brane or free to move in the bulk. With the warped geometry model the extra dimensions are not so large and therefore these interactions are not diluted so much No longer tethered to the weakbrane they could travel anywhere in the bulk and have energies as high as the Planck scale. Only on the weakbrane does the energy have to be less than a Tev. Because these interactions would be in the bulk and therefore operate at high energies, unification of forces would be a possibility. Also the hierarchy problem in warped dimensions requires only that the Higgs particle be on a weakbrane, so that its mass will be about the same as the weak scale energy (the weak gauge bosons need not be stuck there but will automatically have the correct masses). So the weak scale would be protected (at ~1Tev), but unification could still occur at very high energies on the GUT scale, without invoking SUSY (but warped extra dimensions instead) The implications of this is that KK particles five dimensional black holes and strings could be observed at energies accessible to the new particle accelerators (tevatron and LHC) A further development known as RS2 involves only one brane in which the warped bulk space localizes gravity via an exponential fall off. The standard model particles which are on or close to the gravitybrane, only act weakly with the gravity field, which although existing in 5D, behaves (due to the extreme warping of the bulk), like 4D Newtonian gravity. In this highly warped space an infinite extra dimension is possible and yet it would still appear as our 4D world. The main points of this version is that a second weakbrane is found to be unnecessary in determining the correct warping falloff of gravity. Hence we just have a single brane determining the distribution of gravitons which interacts with the Standard Model particles at some distance away from this gravitybrane. So long as this function exponentially reduces the Planck scale (producing the well known high value of the Planck energy that is observed), then we overcome the hierarchy problem. The quantum fluctuations effect for the Higgs particle is reduced, which in turn produces realistic values for the particle masses and their gravitational effects. By employing warped space, it becomes feasible to explain the existence of higher dimensions, without the need to resort to the compactification of the extra dimension (or at least it offers a supplementary method to string theory). When gravity is localized in this way, physical laws behave as if these extra dimensions weren't there, just as with rolled up dimensions. Indeed by considering a negative energy on the gravitybrane (in addition to the warped space due to the negative energy of the bulk space), the brane itself is no longer flat and we can create a locally localized gravity, in which the graviton function only produces a 4D interaction only in a local region of our universe - - -the rest of space could look 5 dimensional or more Matrix theory applies a 2 dimensional matrix property to each point in 10D stacetime and is used to explain the behavior of Do-Branes which move through this space. Even though the theory does not explicitly contain gravity Do-Branes act like gravitons Furthermore the theory of Do-Branes mimics supergravity in 11 dimensions not 10. That is the matrix model looks as if it contains supergravity with one more dimensions than the original theory seems to describe. This has led string theorists to believe that matrix theory is equivalent to M-theory. All this tends to cause us to be cautious when talking about the reality of these extra dimensions. As Ed Witten summarized " space and time may be doomed" and N Seilberg says "I am almost certain that space and time are an illusion" while David Gross states that "Very likely, space and even time have constituents; space and time could turn out to be emergent properties of a very different looking theory" |