(Alex)Dirk Freyling independent researcher & artist

direct e-mail contact Dirk Freyling

 

Knowledge is a privilege. Sharing is a duty. 

Wissen ist ein Privileg. Die Weitergabe eine Pflicht.   

 

[Technical note: In case this web page is unintentionally without a web page menu ("frame-less"), here is the complete elementary body theory web page].

The entire website kinkynature.com is divided "roughly" into two sections and is designed as an information site. Due to the abundance of information and the necessary "comprehension-promoting" readability as well as the aesthetics of the illustrations, a smartphone is not suitable for viewing. The extensive website menu of the science section - on the left and above the selected topic page(s) - requires space. The first choice is a computer monitor. A tablet can be used with some restrictions.

 

The English {Elementarkörpertheorie-Version EKT (elementary body theory EBT)} version has been in the works for a long time, but stil has unfortunately only been realised in fragments to date...

 

The contents and model results regarding the phenomenology, derivation and calculation of concrete quantities, such as the proton radius, anomalous magnetic moments, the derivation of Sommerfeld's fine structure constant, the mass-space couplings that form matter, etc., see selection menu (German) of the ‘frame page’, have been worked out for years.

"NEWS"

A ‘didactically motivated modernisation’ for the purpose of comprehensibility is being considered. To this end, I have put a more interdisciplinary test page online in parallel, which ‘works’ in dialogue with test readers for better comprehensibility. See An overdue rendezvous with causal rationality ...
 

 

What is it all about?

A sustainable paradigm shift.

What is meant here by paradigm shift?

The replacement of the standard (thought) models [SM ΛCDM] of theoretical physics.

[SM] Standard Model of (Elementary) Particle Physics (SM)

[ΛCDM] Standard Cosmological Model (ΛCDM model)

What takes their place?

A much clearer, more effective, interdisciplinary, mathematically simpler model, which describes the micro- and macro-cosmos in a standardised, cross-scale and predictive way.


 

ELEMENTARY BODY THEORY

by  Dirk Freyling       1986 - today

[EBTThe »Elementary Body Theory (EBT)« deals in particular with answers to the question of how mass and space are linked to each other in a fundamental way and how their ‘interplay’ leads to understandable formations of matter that can be calculated formally and analytically - both microscopically and correspondingly macroscopically. Neither a variable time, nor mathematical space-time constructs, nor any form of sub-structuring are demonstrably necessary for a clear understanding and the resulting equations based on the phenomenologically founded elementary body theory.

Less is more...

Contrary to the statement of standard physics which postulates four fundamental forces, elementary (scale-corresponding) mass-space-coupling - reduces any interaction to the mass-to-radius ratio. This leads to constructive, “easy to understand” objects which can be expressed either by its radii or its reciprocally proportional masses.

four fundamental forces   →    mass - space - coupling

 

 

[ Elementary Body Theory ]

I don't like supervised thinking. The extent to which one or the other statement from my fellow human beings corresponds to my ideas is not a criterion for whether I present it.

I am a passionate advocate of aphoristic "language images".

By the way: “The language of good science is bad English.“ This is what one of the most renowned englicist in Germany, Ekkehard König, ironically says.

"Words made here", partly in the form of satire, irony and cynicism, are an expression of indignation and incomprehension about the questionable, established arbitrariness of today's theoretical models for describing matter.

A more precise formulation on my part is often possible, but is then not understood by the majority of readers, due to the strongly mathematised, formal content, since the detailed knowledge that is technically necessary can usually only be understood by those who do not want to hear or read it, since it demonstrably leads the represented analysed standard models ad absurdum.

 

 

Preface

Nature can only add or subtract. A “secured” »higher mathematical reality« exists exclusively within the framework of axiomatically based language (mathematics). To what extent a correct mathematical structure (»higher mathematical reality«) is physically applicable cannot be decided with the “means” of mathematics (see “indisputably exemplary” epicyclic theory and Banach-Tarski paradox). Mathematics ultimately captures quantities and cannot distinguish between vacuum cleaner ("dust cleaner") and dust (Staubsauger und Staub).

If Euclid (...probably lived in the 3rd century B.C.) was still looking for plausible intuition for mathematical foundations and thus made an interdisciplinary connection that could be evaluated as right or wrong, in modern mathematics the question of right or wrong does not arise. Euclid's definitions are explicit, referring to extra-mathematical objects of "pure contemplation" such as points, lines, and surfaces. "A point is what has no width. A line is length without width. A surface is what has only length and width." When David Hilbert (1862 - 1943) axiomatized geometry again in the 20th century, he used only implicit definitions. The objects of geometry were still called "points" and "straight lines" but they were merely elements of not further explicated sets. Hilbert stated that instead of points and straight lines one could always speak of tables and chairs without disturbing the purely logical relationship between these objects.

But to what extent axiomatically based abstractions do couple to real physical objects is another matter altogether. Mathematics does not create (new) phenomenology, even if theoretical physicists like to believe this within the framework of the standard models of cosmology and particle physics.

 

A historical review shows that people who draw attention to mistakes have to reckon with the fact that they will not be believed for decades, while those who publish spectacular “air numbers” within the framework of existing thought models are honored and courted. Critics are perceived by the masses as disagreeable interferers. In the words of Gustave le Bon, “The masses have never thirsted for truth. They turn away from the facts which they dislike and prefer to idolize error when it is able to seduce them. Whoever knows how to deceive them easily becomes their master, whoever tries to enlighten them always becomes their victim. "

Whoever strives to belong to a group does not become a critical thinker. Logical and methodical analytical skills, knowledge of historical relationships, self-developed basic knowledge instead of literature reproduced, ambition, self-discipline and a good portion of egocentrism are basic requirements for independent thinking and acting.

Theory claim and empirical findings

An experiment needs a specific question to be conceived. If the question is the result of a mathematical formalism, the result of the experiment is correspondingly loaded with theory. If the measurable results are then preselected and only indirectly "connected" to the postulated theory objects, as is usual within the standard models, there is nothing to counter the arbitrary interpretation.

Thought models must be absolutely conceptually transformable in order to acquire an epistemological meaning. A mathematical equation that cannot be conveyed outside is always an »epistemological zero« in the context of a physical thought model.

 

 

Unfortunately there is no complete English version/translation for the "Elementarkörpertheorie" yet available. The following explanations provide the basics of the theory and as results important calculations (… Proton radius, ground state energies, calculation of masses (Neutron, Pion(s), Higgs-Boson, ...). You'll find more detailed information if you select certain main issues from the website menu (auf Deutsch). [Technical note: If the present website has no website-navigation-menu, click here, to get the complete frame-based website] "Feel free" to use a common webbrowser translation tool. You'll discover useful information, insights and surprising equations to deduce and calculate physical values based on mass-radius-relations such as... Sommerfeld Fine-structure constant, mass and radius of the universe, Planck units, cosmic microwave background temperature, ...

 

Less is more...

four fundamental forces   →    mass - space - coupling

 

Contrary to the statement of standard physics which postulates four fundamental forces, elementary (scale-corresponding) mass-space-coupling - reduces any interaction to the mass-to-radius ratio. This leads to constructive, “easy to understand” objects which can be expressed either by its radii or its reciprocally proportional masses.

Remember: Divergence problems are theory based. The internal structure of the energy sources are simply not “captured”. Taking into account the finite, real-physical oriented, phenomenological nature of objects, the "infinities" resolve plausibly.

 

Quantum Mechanics on closer inspection

A »time-energy uncertainty (principle)« is often mentioned lightly in connection with the »position-momentum uncertainty principle«. In quantum mechanics, time is not an observable, but a number that parameterizes the chronological sequence of the quantum processes. So there is no* time operator with a unique universal exchange relation that could be investigated... (see for instance *How to Introduce Time Operator Zhi-Yong Wang, Cai-Dong Xiong https://arxiv.org/ftp/quant-ph/papers/0609/0609211.pdf )

Feynman-Stückelberg interpretation for "inexplicable" negative energy values of the Dirac equation.
In the picture of quantum mechanics, this problem was supposedly "solved" with the help of Heisenberg's uncertainty principle, by arbitrarily interpreting the corresponding solutions as entities with positive energy that move backwards in time. The negative sign of energy is transferred to time (Feynman-Stückelberg interpretation). But this is from an epistemologically point of view frankly speaking “nonsense” and leads to a far more controversial aspect, the time-reversible Lagrangian (field theory), which is not observed in real physics. Break your teacup and try to restore it...On closer inspection one has to “bridge” microscopic interactive atoms and macroscopic teacups with a formalizable model, this model would be certainly not based on time reversal. 

The calculation of ground state energies is based neither on quantum mechanics nor on quantum electrodynamics. Because a significantly decisive part is determined by the ratio of the interacting masses. This ratio is neither QM and certainly not QED based (see Bethe-Salpeter equation from Green’s function, Dyson equation). "Look" at the reduced mass term in muonic hydrogen to get an impression how "bad" the situation is. The reduced mass [mred = mA / (1 + mA / mB)] is - whether you want it to be true or not – (historically) derived from "Newtonian celestial mechanics" within the framework of standard physics. In plain language, this means that in terms of atomic interactions ground state energies are neither QM nor QED based.This is one of those touchy issues which are completely out of focus because there is not even a hint how to solve this in the QM framework.

 

 

  [ Elementary Body Theory ]

The self-evident fact that the distance on a spherical surface does not correspond to the "straight" distance between points A and B requires no abstraction.

This leads to the Compton wavelength as a mass equivalent. From this consideration follow "geometrically" mass-specific radii. Compton wavelengths are measurands. [ CODATAC(Proton)  CODATAC(Elektron) ]

 

 

"abstract"

The Elementary body theory (Elementarkörpertheorie [EKT] ) is based on plausibility and minimalism and provides phenomenologically based equations without free parameters and a formalism which leads to results which are in good to very good agreement with experimental measured values. For a clear understanding and as a result of the phenomenologically based elementary particle theory generated equations, neither a variable time, nor mathematical space-time constructs, nor any form of substructuring are necessary.

The time-dependent elementary body equations are derived from the observed invariance of the (vacuum) speed of light. The fundamental difference to the (special) theory of relativity respectively to the Lorentz transformation is the radially symmetric-dynamic character of these equations.

The main object of the elementary body theory is the elementary body originally a pulsating hollow sphere. At maximum expansion the hollow spherical shell-mass is at rest. The equations of motion - based on a sine function - describe the complete transformation of motion energy without rest mass (photon) to mass.

The basic mass-space-model requires that the equations portray both the massless photon and mass. The equations r (t) = r0 · sin (c · t / r0) and m (t) = m0 · sin (c · t / r0) do exactly that. The timeless speed of light - as a state of pure motion - is not contradictory with the matter-energy-embodiment.

 

The transformation from a photon to a mass-radius-coupled space does not correspond phenomenologically to a partial oscillation, as was initially assumed (also) within the framework of the elementary-body-model. The matter-forming transformation of a photon corresponds to an irreversible »state change«.

 

 

Time reversal, as required "mechanistically" from classical physics to quantum mechanics, is in general contradictory to measuring reality (thermodynamic processes). The fully developed elementary body (r (t) = r0, m (t) = m0) can not regain the state of the photon by itself.

The time-dependent mass formation is coupled to the time-dependent radius magnification r = r (v (t)). In simple words, the initial, pure motion energy gives rise to time-dependent spherical surfaces, which as such span a space whose reciprocal size is a measure of the equivalent mass. After a quarter period (½ · π), the elementary body is fully developed (r (t) = r0, m (t) = m0), meaning that the expansion velocity v (t) is zero.

Since the process of resting-mass reduction corresponds to an inversion of the relativistic dynamics of a velocity-dependent momentum mass, the internal dynamics for energy conservation of the elementary body is suggestively called momentum-mass inversion.

 

 

State as information = photon

t = 0 ,  the entire energy is available as pure information, mass- and spaceless

 

Information as a material condition = elementary body

t (½π) ,  the total energy is "present as" mass m0 with radius r0

Phenomenologically, the transformation of motion information into spatial information is complete. Without external interaction the elementary body remains in this state. If the elementary body is "excited" from the outside, different interaction scenarios occur which, depending on the energy of the interaction partners, lead to partial annihilation or (full) annihilation. Matter-forming partial annihilations are formed in the simplest form by the proton-electron interaction (keywords: Rydberg energy, hydrogen spectrum). Mass-coupled space annihilates according to r (t) and m (t). "Radiation" is absorbed or emitted. The interaction reversibility that is possible must be via excitation from the outside. This could be the interaction with other elementary bodies, photons or "embodied fields", which can always be understood as elementary body (states).

As mentioned in the context of the derivation of the mass-energy equivalence E = mc², the basic misunderstanding ("outside" the elementary-body theory) is that the properties of an interacting photon are projected onto the "resting state" of the photon. However, according to equation [P2.3] and its temporal derivative [P2.3b], as well as [P2m], the »resting state« of the photon is the space- and massless, "light-fast" (energy) state of maximum motion. This means that an information is propagated that "unfolds" only upon absorption (interaction) of the photon in accordance with equations [P2.3], [P2m] and their derivatives, and then the time-dependent phenomena of interference and (mass-based) collision shows. In regard to photons in interstellar space, the light path and thus the photon is invisible. Only when an interaction (absorption) "appears", the photon becomes visible (detectable).

 

 

Exact calculation of  the Proton radius

A consideration of the elementary body provides an accurate theoretical value for the proton radius. Elementary body theory based the proton radius is the Compton wavelength λC of the proton divided by π/2.

This result is in excellent accordance with the measured value of the proton radius (investigation muonic hydrogen, July 2010 and 2012/2013 at the Paul Scherrer Institute in Switzerland.

                                                                                                                                                                                                                                                "abstract"

 

[ http://www.psi.ch/media/proton-size-puzzle-reinforced ]

 

 

All cosmological "observational studies" are not controllable laboratory experiments. The postulated theoretical implications strongly influence the experimental interpretations. The human observation period is extremely small compared to the periods of time in which cosmic movements took place and played out. To substantiate assumptions with the data from the human observation period is "far-fetched" to put it casually. All current supposedly empirical measurements are (big bang) theory laden. Postulated time spans, distances, and energy densities are subjective-theoretic. The entire present physical view is based on the paradigm of "physical space-time".

Already Isaac Newton thought that the idea that gravitation could work through empty space was absurd. It is - superordinate and considered as a whole - anything but trivial to regard space and time as physical "objects". Space and time are primarily "order patterns of the mind". In order to "preserve" physics from these mind-patterns, a phenomenological examination and explanation are absolutely necessary.

By the way...The earth does not move around the sun. As Isaac Newton already stated in his Principa 1687, the sun and planets move around the centre of mass of the solar system, called the barycentre.

“Remember”: Spacetime is not measurable. The postulate states that within the framework of the (formalised) consideration of the theory of relativity, time and space only exist as a unity. One consequence: we cannot record events in the postulated general relativistic cosmos as a „snapshot” in time, because time does not exist in space-time. According to the theory of general relativity (GR), there is only spacetime and spacetime cannot be experienced by the senses or by measurement.

By the way...Popular scientific explanations of the General Theory of Relativity (GR) are all inadmissible interpretations, because the complexity necessary for the understanding is not considered. This is as if somebody presents Chinese characters to an audience not competent in writing for viewing and discussion. Then everything and nothing can be interpreted into it, since nobody has the conditions for the decoding as well as nobody the ability of creating Chinese characters. To formulate it in a bold way: In GR, also writing errors lead to new solutions and this is (already) true for people who do differential geometry professionally.

Despite the impossibility of depicting various mathematically generated theoretical objects and interaction scenarios in a generally understandable and plausible way, it is didactically widespread to convey knowledge with "simplified" (unfortunately wrong) views. An example is the balloon inflated with markings (which are supposed to represent planets, stars, galaxies), which is supposed to depict the cosmic expansion after the Big Bang. This creates a sense of vivid understanding in the viewer. Only this balloon analogy is fundamentally wrong. Because…

space or, in simplified form, a balloon surface that expands, but the mathematical construct of space-time.

iii) According to today's ΛCDM-model ideas the expansion is accelerated. In the "balloon understanding", the radius expands uniformly or at a slower rate (up to a material-related maximum before the balloon bursts).

iv) Also concealed are the (faster-than-light, strongly limited in time) postulated, theory-dependent inflation-time-period as well as the assumptions and effects of the postulated dark matter and dark energy, which do not occur in the balloon model.

And overall the arbitrariness problem of the free parameters is compounded by the unavoidable "axiomatic violation" of the covariance principle. The general theory of relativity was born, among other things, out of the demand to be able to use any coordinate system to describe the laws of nature. According to the covariance principle, the form of the laws of nature should not depend decisively on the choice of the special coordinate system. This requirement is causally mathematical and leads to a variety of possible coordinate systems [metrics].

The systems of equations (Einstein, Friedmann) of the General Theory of Relativity, which form the basis of the statements of the Standard Model of Cosmology, do not provide analytical solutions.

According to GR postulate, not only mass but also every form of energy contributes to the curvature of space-time. This applies including the energy associated with gravity itself. Therefore, Einstein's field equations are non-linear.

But only idealisations and approximations lead to computable solutions to a limited extent. The unavoidable ("covariant") contradictions come with the obviously inadmissible idealisations and approximations of the system of nonlinear, chained differential equations. Mathematically, the covariance principle cannot be "violated" because it is axiomatically justified. Only this axiomatic premise "disappears" with the mutilation" (idealisations and approximations) of the actual equations. In other words, the mathematically correct equations have no analytical solutions. The reduced equations (approximations, idealisation) do have solutions, but these are not covariant. Thus, no solution has a real-physics-based meaning. This way of using mathematics is arbitrary, since different results are obtained depending on the "taste" of the (self-)chosen metric.

The 'big bang problem' as a 'primordial singularity' only occurs theoretically if Einstein's credo (: irreversibility is an illusion) is true. This conclusion means, however, that thermodynamics is consistently negated from the outset and one restricts oneself to the flawless Hamiltonian mechanics as the basis of GR, or at least only isentropic processes (no entropy change) are considered in order to at least 'save' the phenomenon of background radiation...

 

 

..the mirage: the connection between information and entropy. Two words, which - according to John von Neumann - both cannot be understood, are linked by a logarithm and a constant that provides dimensions. What is even more astonishing: Many cosmologists believe in all seriousness that this will bring thermodynamics and its entire conceptual apparatus on board. Hence Hawking's 'logic': “The problem in Bekenstein's argument was that a black hole, if it had a finite entropy proportional to the area of its event horizon, would also have to have a finite temperature. It would follow that a black hole could be in equilibrium with thermal radiation at any temperature other than zero. But in classical terms, such a balance is not possible because the black hole would absorb without emitting. What the international elite of theoretical cosmologists never took notice of was the simple fact that 'their entropy' is more of a 'Shannonian entropy', namely information regarding however, no thermodynamic entropy. In addition, the main principles of thermodynamics and the equations of motion derived from them are formulated with process variables that locally have nothing to do with the Einstein geometry of GTR. This immediately leads to the problem of the extent to which the Second Law is compatible with the ART at all. If one absolutely wants an 'informal temperature' conjugated to information, one has to derive it from 'information theory'; but it certainly has nothing to do with the Kelvin temperature and therefore nothing at all with physical radiation processes. The 'big bang problem' as a 'primordial singularity' occurs theoretically only if Einstein's credo (: irreversibility is an illusion) applies. However, this conclusion then means that one consistently negates thermodynamics from the start and restricts oneself to flawless Hamilton mechanics as the basis of GTR, or at least only envisages isentropic processes (no change in entropy) in order to at least 'save the phenomenon of background radiation '.

 

The “secret” of very weak gravitation in relation to the electrical interaction and strong interaction is based on the false assumption that there is generally a mass decoupled space existing. If one considers inherent space that elementary particles and macroscopic bodies contain, by their object expansions and by their radius of interactions, then it becomes clear that the "missing" energy is (in) the space itself. To create space(-energy) space-coupled mass(-energy) must be transformed in such. One consequence, inflation is ruled out. Another consequence of the observed “real object physics” is, the “massless concept” of the SM is ruled out.

The gravitational constant γG used in the "known" Newton's law of gravitation refers to the "length-smallest" body G {elementary quantum}. This is not obvious since the "normal formulated" law of gravity does not explicitly disclose this original connection.

The secret of very weak gravitation in relation to the electrical interaction and strong interaction is based on the false assumption that there is generally a mass decoupled space. If one considers the space that the macroscopic body spans both by its object expansion and by its radius of interaction, then it becomes clear that the "missing" energy is (in) the space itself. In this sense, the gravitational constant γG is the "measure of things" for macroscopic bodies.

 

Macroscopic bodies and gravity

For bodies with radius-mass ratios different from rG / mG, this means „colloquially simply” that "work" had to be done to span a larger (body) space than is naturally-coded in the smallest, most massive elementary quantum {G}. Taking energy conservation into account, this energy can come only from the mass-dependent resting energy. In the mass-dependent interaction of gravitation, only the mass fraction (effective mass), which is available after deduction of the mass of equivalent space energy, is then carried.

Well-known macroscopic objects (... billiard ball, football, earth, sun, ...) obviously do not satisfy the mass-radius-constant equation [F1]. Its real size is larger by many orders of magnitude (even before the interaction) than Equation [F1] demands mass-radius-coupled for elementary bodies.

In addition, the ratio (RO / MO) is many orders of magnitude greater than the governing elementary quantum. Without knowing the concrete nature of many-body interleaving, it can be generally understood that the [space] energy of the gravitational interaction, which seems to be missing in relation to the rest energy, lies in the real physical object expansion, which is determined by the object radius r0 and by the Interaction distance r (interaction radius).

 

Space energy calculation :

Practical definition

Assuming that the elementary quantum {G} is the smallest in length and inherently most massive body, it follows that the space energy zero point is present here. This means that all values for the space energy ER equation [ER] are additive relative to this zero point of the space energy.

 

 

 

From this follows the effective interaction mass Meff = (rG / r) · mx, which is indirectly expressed by the gravitational constant in Newton's law of gravitation. The gravitational self-energy equivalent to the effective interaction mass can be derived (also) from the radius-mass-coupled energy of the elementary body (compare equation [E1r]), taking into account the "scaling" using rG / RO.

It follows from these simple plausibility considerations that gravitation, on the basis of the phenomenology of an energy-conserving, mass-radius-coupled space, can be formally analytically determined within the framework of „simplest” mathematics. The inner spatial composition and interleaving of the atomic or molecular structure of the macroscopic many-body objects has no influence on the gravitational force or gravitational energy as long as the interaction radius r is greater than the object radius RO (r > RO = "elastic interaction").  

 

 

Self-similarity

The assumption is that the ratio of time-dependent universe radius to time-dependent universe mass is time-independent(!). The multiplication of (rUni / mUni) with c² is equal to the gravitational constant γG. These assumptions include a "beautiful" correspondence between cosmos, gravitational constant and the elementary quantum {: G with rG, mG} respectively the Planck sizes for mass (mPl = ½ · mG) and length rPL (radius, mPl= ½ · rG). These assumptions lead to concrete calculations, such as the maximum universe mass, the maximum universe radius and with the help of the hydrogen parameters (electron, proton mass and ground state energy) to the calculation of the background radiation temperature.

 

Ernst Mach (1838 - 1916) noted that the "weakness" of gravitation is related to the "enormous" size of the universe. He suspected a relation between universe mass, universe radius and gravitational constant. Sir Arthur Stanley Eddington (1882-1944) obtained the ratio of universe radius to universe mass by dividing the gravitational constant by c² (... formally equal in the context of the correspondence described here between "length lesser" and lenght-largest And Erwin Schrödinger had already expressed this as a possibility as early as 1925. Source: Mach's Principle - Relating Gravity to the Universe

Why this simple assumption did not prevail has most likely to do with the (later) favored (inflationary) big bang theory. There was no room for "(corresponding) simple thoughts".

 

elementary body based universe

(more) details, phenomenology, ... see here

 

Hydrogen is by far the most abundant matter of the universe. Hydrogen accounts for approximately 90% of interstellar matter. As shown, the omnipresent hydrogen in the universe is the "source" of  3K background radiation.

For reasons of consistency, the Rydberg energy value results stringently from the elementary-body theory-based proton-electron interaction. This ensures that all the equations used can be linked together without any approximation. The deviation from the spectroscopically measured Rydberg energy (ERyexperimental) is: ERyEB-theory/ERyexperimental ≈ 1.0000025.

 

Conclusion: The above basic analysis of gravity disenchanted various myths (keywords: graviton, inflation field, dark energy, dark matter, ...) to form a universe. Neither "ART-standard" differential-geometric considerations, super-light velocity or four-dimensional space-time constructions are necessary. Furthermore, the gravitational interaction described here can not be phenomenologically "unified" with quantum field theory (QFT), since there is no need to do so. The postulated "exchange particle objects" of the Standard Model of Particle Physics (SM), ie gluons and vector bosons and, in the broadest sense, neutrinos, do not "couple" to gravity even if these "force mediators" would exist. Without explicitly stating this, however, there are neither gluons, vector bosons nor neutrinos (see the chapters Standard Model, Neutrinos and the remarks (further down) about the Higgs-Boson and SM-fantasies).

 

 

 

Furthermore:

Concept of electric charge  

Electric charge is a secondary term/concept of standard phycics that suggests a "phenomenological entity" that is uncoupled from the mass (and the radius) of the charge carrier.

Based on elementary-body theory all charge interactions are clearly traceable to mass-radius couplings. Conveniently, electrical charges in the elementary-body model occur only as an implicit function of the Sommerfeld fine-structure constant α as a (formal) result of the mass-radius coupling.

"Keys" for understanding the formation of matter are the phenomenologically founded charge possibilities. First, the energetically (strong) elementary body charge q0 (which energetically equals m0) and the elementary electric charge e.

 

f7 was "introduced" to show that the [elementary body] charge q0 is ("only") a scaled mass-radius function.

 

 

Side note

Particle physicists generally use the phenomenologically incorrect term decay even though they mean transformation. Decay would mean that the decay products were (all) components of the original particle. But that is not the case, at least not within the theoretical implications and postulates of the Standard Model of particle physics (SM).

 

For convenience, "powers of ten" are written as follows: 1.23 · 10-3 = 1.23e-3. This has the advantage that the interested reader can take over the corresponding numerical values in the "computer" for "control" (copy & paste).

The uncertainties of the (CODATA) (mass) values and the natural constant values are not explicitly named. Example: Mass of the electron: 9,10938356e-31 kg 9,10938356 (11) e-31, means a standard deviation of ± 0,00000011e-31 kg. It is calculated here with the value 9,10938356e-31 kg.

 

Charge-dependent matter formations

Basics

The extended charge principle leads beyond elementary-body theory-based hydrogen atom-forming to additional proton-electron interactions. From the generalized, clear phenomenological process stringently follow the neutron (e-q0 interaction) and pions (q0-q0 interaction) as energetically possible (time-instable) "particles". Without concretizing this here, the charged pions “decay” (convert) into muon and anti-muon and then into electron and positron. Overall, "diverse elementary particles" can be formed in the context of the extended charge concept in "formal analogy". Noteworthy is the fact that this formalism provides simple, without free parameters solutions that are in good agreement with the (energy and mass) values of the "formed particles".

On the basis of the reduced mass of the electron it can be easily shown how a model view "works wrongly". Considering celestial mechanics, a "small" centroid shift results from the proton to the electron, since the mass of the proton is finite. From the point of view of two equal charges, this assumption is unfounded, since masses in the standard view of physics only have an effect via the (misunderstood) gravitation, which is smaller by almost forty powers of ten to the electrical force. Overall, in the world view of the prevailing physics, a mass can not interact with a charge, because there exists simply no such phenomenology. It is astonishing how this fact was ignored mass psychologically over generations and is still ignored. 

 

... there is no mass-center-shift "between two equal I charges I

 

BUT : The model view that the interaction between "charges" that are at a distance r from each other does not occur in matter-forming elementary-body theory. In elementary-body theory, the "charges" overlap with/in a common origin.

                                                                                                            

                             

Reduced mass alternative in a mass-space-coupled model

Let's start with the superposition of two elementary bodies A and B, with the masses mA and mB and the mass-coupled radii rA and rB. By mass-radius-constant equation [F1] there is no "room" for interpretations ► The result applies to all charge carrier constellations (... A-B, proton-electron, proton-muon, ...)

 

Here one can clearly see that the alleged center-of-gravity correction of the "celestial-mechanical model has nothing to do with  (for example proton and electron) with focus on two interacting charges (at a distance r), since electron and proton, as equal charges, can not undergo a shift, either phenomenologically or computationally.

NOTICE! Equating an electrical centripetal force with a (only) mass-dependent centrifugal force is phenomenologically unfounded in the context of physics and is reminiscent of the epizykel theory. The expression for the resulting mass m(rA + rB) in equation [MAB] is mathematically identical to the celestial-mechanical centroid correction of two macroscopic masses (reduced mass) which computationally interact elastically as point masses, but the phenomenology for the equation [MAB] is a completely different one.

Furthermore, the calculation of ground state energies is neither quantum mechanically nor quantum electrodynamically possible. Since a significant amount is determined by the ratio of the interacting masses. There is neither QM nor QED based the possibility of introducing the reduced mass mred = mA / (1 + mA / mB) quantum-field-phenomenologically. The reduced mass - whether one wants it to be true or not - is historically derived from "Newtonian celestial mechanics" within the framework of standard physics. This means in plain language that in terms of atomic interactions, these are neither QM nor QED justified. QM and QED are "epicyclic".

 

 

 

Mass - inherent radius (of the electron) re

The SM-based postulated theoretical implications strongly influence the experimental interpretations of the SM protagonists ... The resulting complex expectation construct leads to a series of fatal misunderstandings. By way of example, the "typical" energy-dependent electron radii, less than 10-19 [m], in connection with particle accelerators are not "false" (measured), but the inferences concerning resting electrons resulting from elementary particle physics are false. The elementary-body theory (on the other hand) consistently describes both the behavior of the "conventional" scattering energies of the scattering partners (of the electron) and of high-energy particles in accelerators.

 

[Practical note: If you can't "follow" respectively won't accept logic based facts, consider the "following"... 

Since almost all successfully SM-indoctrinated people get a crisis by „hearing“ the term electron radius and "instantaneously" think of the unpopular classical electron radius, they should “replace” the electron mass inherent radius (re) through the "problem-free" Compton wavelength of the electron (λe) and postpone their "communication problem". Don't dramatize the model view : λe = ½ · π  · re]

 

 

First of all, the differential scattering cross section dσ/dΩ always comprise the form : dσ/dΩ = (e²/me4πε0c²)² · fW for electrons in "conventional" scattering experiments.

fW indicats the specific type of scattering. The „communication problem” of the prevailing physics with respect to the (unpopular) classical electron radius re(classical) is resolved by analyzing the recurrent mass-radius-coupled term (e²/me4πε0c²) energetically. Finally, the term (e²/me4πε0c²) consists of the ratio: electrical energy to total energy (α/4) and the electron mass inherent electron radius re [ = (dσ/dΩ) / ((α/4) · fW)], multiplied by π/2 corresponds to the Compton wavelength of the electron (λe = ½ · π  · re). 

Furthermore, the object radius changes with the velocity radially symmetric and not just one-dimensional in the direction of movement. Here, too, elementary-body theory provides a phenomenologically-based, consistent, realobject-oriented model. 

One can not mention it often enough: Apart from "SM-based interpretations" of particle accelerator physics, the electron mass inherent electron radius re respectively the classical electron radius re(classical) = ((α/4) · re), is used in all (!) equations for the calculation of the scattering cross sections for elastic and inelastic scattering of electrons (keywords: Møller scattering, Compton scattering, electron-positron pair production, photoelectric effect, Klein-Nishina equation, Bethe-Bloch-Sternheimer equation, Kramers-Heisenberg formula, etc .). re has a phenomenologically based connection with the Compton wavelength (of the electron). Here, so to speak, "avenges" the lack of a consistent, phenomenologically motivated (standard physics) concepts of electric charge. See the various mathematical approaches of classical electrodynamics, quantum electrodynamics (QED), quantum chromodynamics (QCD) and in particular the electroweak theory. In the context of elementary-body theory, the electric elementary charge (e) and the associated electrical energy E(e) expresses a mass-radius-coupled energetic ratio. For this »understanding problem« exists the biggest emotional outbursts and a methodical plausibility refusal from standard physicists and their followers (... more details see the chapter electron radius).

 

 

 

Charge-dependent matter formation possibilities 

The charge-dependent matter formation generally describes the A-B interaction possibilities. A and B are elementary bodies with the masses mA and mB and the reciprocal proportional radii rA and rB. The following applies: mA · rA = mB · rB = FEK = 2h / πc  [F1]. The phenomenologically founded formalism leads to the equations:

In the above (matter-forming) α-function-equations, only the masses mA and mB of the interacting elementary bodies occur as variables. The charge as such, or more precisely the charge size, is implicitly determined by the functional relationship of the Sommerfeld fine structure constant α. (Details and derivations see »Ladungsabhängige Materiebildungen«) 

 

e-e interaction

The term e-e interaction means that two elementary charge carriers interact. The most prominent example of this type of interaction is the proton-electron-based hydrogen atom.  

mA  = me = 9,10938356e-31  kg : electron-mass        mB  = mp = 1,672621898e-27  kg  : proton-mass

c =  2,99792458e+08 m/s  α = 0,0072973525664

 

e-q0 interaction

elementary body carrier A(q0) interacts with elementary carrier B(e). The most prominent example of this type of interaction is the proton-electron-based neutron.

The neutron mass mn arises from a matter-forming charge interaction of the electron and proton and can be understood and calculated by the interaction of the elementary body charge q0 for the electron and the elementary electric charge e for the proton.

me = 9,10938356e-31  kg        me(q0) =   (4/α) · me =  4,99325391071e-28 kg        c = 2,99792458e+08 m/s

mp 1,672621898e-27 kg

∆m = 1,405600680072e-30 kg    ∆Eee =    1,263290890450e-13 J      ~    0,78848416 MeV

Taking into account the phenomenologically-based, approximation-free approach, in formal-analytic form of the equation: mn = mp + me + Δm [mq0e], the "theoretical" result of elementary particle theory based neutron mass calculation (according to charge-dependent proton-electron interaction) is “sensational”. In addition you'll find below a calculation of the magnetic moment of the neutron  (see SUMMARY of FORMULAS).

 

 

q0-q0 interaction

Charge carriers A and B interact via the elementary body charge q0. The most prominent example of this type of interaction is the proton-electron-based charged pion.

 

                                           

                        mA  = me = 9,10938356e-31  kg : electron-mass

                              q0mA =   (4/α) · me =  4,99325391071e-28 kg

                                   mB  = mp = 1,672621898e-27  kg  : proton-mass

                             q0mB =   (4/α) · mp =  9,16837651891e-25 kg

                                 c = 2,99792458e+08 m/s      α = 0,0072973525664

                                           [mq0q0]

                                  ∆m = 4,99053598e-28  kg     (∆m/2) / mπ(exp)  ~  1,00289525

                              ... ∆m means the mass of two charged pions (matter creation)

 

The extent to which experimental particle physics can accurately determine resting pion masses is highly doubted. The neutral pion is a "pion" due to the different mass of the charged pions only in the SM requirement. The abstraction, which is "equal" to particles with different masses according to postulated QM superpositions (keyword: quarkonia), is one of the many arbitrariness hypotheses within SM (see SM-quark mass uncertainty in the percent error range) and "outside" of mathematical formalism of the SM unfounded.

 

For mass-like interaction partners (for example, proton-antiproton or electron-positron) the general α-function equations simplify - for example the  q0-q0 interaction - to the equations ([Eq0q0] and [mq0q0]):

 

"Surprising" is the "circumstance" that in the context of "charge-dependent matter formation the strong proton-antiproton interaction follows a matter-formation energy of ~ 257 GeV depending on the (anti) proton mass and the Sommerfeld fine-structure constant α which, according to charge conservation, produces as a variation possibility two mass-radius-coupled "small mass heaps" (Masse-Häufchen) which map uncharged and charged Higgs-boson masses. The final equation [2q0q0] is unbeatable in terms of simplicity.

 

                                                      α  = 0,0072973525664

                                            m := mp = 1,672621898e-27  kg  : (anti-)proton-mass

                                                         ∆E( p+, p- )  =  257,15410429801 GeV

                                                       ∆m( p+, p- )   =   4,584188259456e-25   [kg]  [2q0q0]

                                                  (∆m( p+, p- ) / 2) = 128,57705215 GeV/c²

                                                     mH(0)  ~ 2,228e-25 kg   ~  125 GeV/c²

                                                   (∆m( p+, p- ) / 2) / mH(0)  1,02861642

This means that with an "error" ~ 2.9%, based on the Higgs boson mass "detected" at the LHC (mH(0) ~ 125 GeV / c²), elementary particle theory predicts an event which exists in the standard model of Particle physics (SM) only as a theoretically predetermined methodical circular conclusion.

 

by the way ...

There are aspects of the Higgs-Boson-mass predictions which are barely known. 

 

David and Sidney Kahana's predictions about the Higgs-Boson-mass and the Top-Quark-mass (1993!!!) in a “parameter free fashion” are very precise. Source: https://arxiv.org/pdf/hep-ph/9312316.pdf

 

According to the standard model (SM) predictions are not possible. How do you explain the obvious discrepancy?

 

Peter Higgs knew about their work … he said, “You’re from Brookhaven, right. Make sure to tell Sid Kahana that he was right about the top quark 175 GeV and the Higgs boson 125 GeV” [Kahana and Kahana 1993].”… Source:https://arxiv.org/pdf/1608.06934.pdf

 

One would assume that highly accurate calculations about the Top-quark-mass and the Higgs-mass are remarkable. Why didn’t the “Kahanas” get the “proper” attention? Why is there no adequate mention about these theoretical achievements?

 

I strongly believe that Sidney and David Kahana’s predictions need to be published and discussed again. With reference to Nambu Y and Jona-Lasinio G 1961 Phys. Rev. 122 (1) 345-358 (https://journals.aps.org/pr/pdf/10.1103/PhysRev.122.345)  it seems that the entire SM-project had already been completed (…“methodical circular conclusions") at the beginning of the 1960s.

 

For further reading see https://arxiv.org/pdf/1112.2794.pdf ... "predictions by the authors D. E. Kahana and S. H. Kahana , mH = 125 GeV/c² uses dynamical symmetry breaking with the Higgs being a deeply bound state of two top quarks. At the same time (1993) this model predicted two years prior to the discovery to the top its mass to be mt = 175 GeV/c²..."

Notice! There is just one outstanding »prediction paper« (1993 https://arxiv.org/pdf/hep-ph/9312316.pdf) which leads to the Higgs-Boson-mass and the Top-Quark-mass with the same theoretical approach prior to the experimental confirmation in 1995 (Top) and 2012 (Higgs).

 

 

 

The uncharged pion ... a pion matter possibility from q0 - q0 interaction

mA = mB = me± = 9,10938356e-31  kg : electron-mass = positron-mass

                        q0mA =   (4/α) · me =  4,99325391071e-28 kg

 

                   

                

                               ∆E( e+, e- )    =      140,05050232093 MeV    [E2q0q0]

                         ∆m( e+, e- )         =   2,496626955355e-28   kg     [2q0q0]

                                         2,4061764315e-28 kg      mπ0  SM - theory laden

                                                     ∆m( e+, e- ) / mπ0  ~ 1,037591

 

 

 

SM linguistic usage…

Particle physicists generally use the phenomenologically wrong term decay although they mean transformations. Decay would mean that the decay products were (all) components of the decaying. But this is not the case, at least not within the framework of the theoretical implications and postulates of the Standard Model of particle physics.

 

                                                                 Wishful thinking and reality

I find it quite amusing and right to the point how Claes Johnson, a Professor of Applied Mathematics, classifies ...

Claes Johnson about QM and SRT

Concerning the crisis of modern physics it is commonly accepted that one reason is that the two basic building blocks, relativity theory and quantum mechanics, are contradictory/incompatile. But two theories which are physical cannot be contradictory, because physics which exists cannot be contradictory. But unphysical theories may well be contradictory, as ghosts can have contradictory qualities.

The Special Theory of Relativity of Einstein is unphysical because the Lorentz transformation is not a transformation between physical coordinates, as strongly underlined by its inventor Lorentz, but misunderstood by the patent clerk Einstein believing that the transformed time is real and thus that time is relative. Quantum Mechanics is unphysical because its interpretation is statistical which makes it non-physical, because physics is not an insurance company. Here Einstein was right understanding that God does not play dice.

Professor of Applied Mathematics, Royal Institute of Technology (KTH) Stockholm , Sweden

Claes Johnson Blog

______________________________________________

 

 

Albert Einstein

It could be helpful to remember what Albert Einstein wrote on quantum mechanics: [1] "The ψ function is to be understood as a description not of a single system but of a system community [Systemgemeinschaft]. Expressed in raw terms, this is the result: In the statistical interpretation, there is no complete description of the individual system. Cautiously one can say this: The attempt to understand the quantum theoretical description of the individual systems leads to unnatural theoretical interpretations, which immediately become unnecessary if one accepts the view that the description refers to the system as a whole and not to the individual system. The whole approach to avoid 'physical-real' becomes superfluous. [Es wird dann der ganze Eiertanz zur Vermeidung des ‘Physikalisch-Realen’ überflüssig.]  However, there is a simple physiological reason why this obvious interpretation is avoided. If statistical quantum theory does not pretend to describe completely the individual system (and its temporal sequence), then it seems inevitable to look elsewhere for a complete description of the individual system. It would be clear from the start that the elements of such a description within the conceptual scheme of the statistical quantum theory would not be included. With this, one would admit that in principle this scheme can not serve as the basis of theoretical physics.”

[1] A. Einstein, Out of my later years. Phil Lib. New York 1950 Seite 498

According to the Copenhagen interpretation of 1927, the probability character of quantum theoretical predictions is not an expression of the imperfection of the theory, but of the essentially indeterministic (unpredictable) character of quantum physical natural processes. Furthermore, the "objects of formalism" "replace" reality without possessing a reality of its own.

The Copenhagen interpretation is characterized by the „comfort” it provides to its "believers”. The wave-particle dualism allows a "transition" to the "wave" with an e-function with complex exponent, which according to Fourier theorem in turn allows "ALL" piecewise monotonous, so also any experimental result to represent formally mathematical. The statistical interpretation refrains from the effort to explore the physical process, because in principle only probability statements in quantum processes are possible

In the time after the Second World War, the Copenhagen interpretation had prevailed, in textbooks was now only the Heisenberg-Bohr quantum theory without critical comments to find.

 

___________________________________________

 

Brigitte Falkenburg writes in Particle Metaphysics: A Critical Account of Subatomic Reality (2007) among other things…

"It must be made transparent step by step what physicists themselves consider to be the empirical basis for current knowledge of particle physics. And it must be transparent what the mean in detail when the talk about subatomic particles and fields. The continued use of these terms in quantum physics gives rise to serious semantic problems. Modern particle physics is indeed the hardest case for incommensurability in Kuhn’s sense."… Kuhn 1962. 1970

…"After all, theory-ladenness is a bad criterion for making the distinction between safe background knowledge and uncertain assumptions or hypotheses." 

… "Subatomic structure does not really exist per se. It is only exhibited in a scattering experiment of a given energy, that is, due to an interaction. The higher the energy transfer during the interaction, the smaller the measured structures. In addition, according to the laws of quantum field theory at very high scattering energies, new structures arise. Quantum chromodynamics (i.e. the quantum field theory of strong interactions) tells us that the higher the scattering energy, the more quark antiquark pairs and gluons are created inside the nucleon. According to the model of scattering in this domain, this give rise once again to scaling violations which have indeed observed.44 This sheds new light on Eddington’s old question on whether the experimental method gives rise to discovery or manufacture. Does the interaction at a certain scattering energy reveal the measured structures or does it generate them?"

44 Perkins 2000, 154; Povh et al 1999, 107 – 111

…"It is not possible to trace a measured cross-section back to its individual cause. No causal story relates a measured form factor or structure function to its cause"…

…"With the beams generated in particle accelerators, one can neither look into the atom, nor see subatomic structures, nor observe pointlike structures inside the nucleon. Such talk is metaphorical. The only thing a particle makes visible is the macroscopic structure of the target"…  

…"Niels Bohr’s quantum philosophy…Bohr’s claim was that the classical language is indispensable. This has remained valid up to the present day. At the individual level of clicks in particle detectors and particle tracks on photographs, all measurements results have to expressed in classical terms. Indeed, the use of the familiar physical quantities of length, time, mass and momentum-energy at a subatomic scale is due to an extrapolation of the language of classical physics to the non-classical domain."...

 

The Quark Parton Model (QPM), developed by Richard Feynman in the 1960s, describes nucleons as the composition of basic point-like components that Feynman partons called. These components were then identified with the quarks, postulated by Gell-Mann and Zweig at the same time a few years earlier. According to the Quark-Parton Model, a deep inelastic scattering event (DIS deep inelastic scattering) is to be understood as an incoherent superposition of elastic lepton-particle scattering processes.

A cascade of interaction conjectures, approximations, corrections, and additional theoretical objects subsequently "refined" the theoretical nucleon model.

A fundamental (epistemological) problem is immediately recognizable. All experimental setups, implementations, and interpretations of deep elastic scattering are extremely theory based.

Fundamental contradictions exist at the theoretical basis of the Standard Model of particle physics, which, despite better knowledge, are not corrected. An example:

The nonexistent spin of quarks and gluons

A landmark, far-reaching wrong decision was made in 1988.

The first assumption was, due to the theoretical specifications of the mid-1960s that in the image of the SM the postulated proton spin is composed to 100% of the spin components of the quarks. This assumption was not confirmed in 1988 in the EMC experiments. On the contrary, much smaller, even zero-compatible components were measured (ΔΣ = 0.12 ± 0.17 European Muon Collaboration). Also the next assumption that (postulated) gluons contribute to the proton spin did not yield the desired result. In the third, current version of the theory, quarks, gluons (...virtual Quark-anti-Quark pairs if one wishes too) and ... their dynamical-relativistic orbital angular momentum generate the proton spin.

 

On closer inspection, the second readjustment has the „putative advantage” that the result in the context of lattice gauge theory and constructs, such as "pion clouds", algorithmically "calculated", can’t be falsified. But this purely theoretical based construction obviously does not justify the classification of quarks as fermions. No matter how the asymmetrical ensemble of unobservable postulated theoretical objects and interactions is advertised and will be advertised in the future, the quarks themselves were never "measured" as spin-½ particles.

 

Summary in simple words: It is possible to create a theory-laden ensemble of Quarks and “other” theory objects and their postulated interactions, but the Quark itself - as an entity - has still no intrinsic spin -½ in this composition. That means that Quarks aren’t fermions, no matter what the actual theoretical approach would be! This is a basic, pure analytical and logical statement.

 

Generally speaking: If one postulates a theoretical entity with an intrinsic value but one discovers that one needs to add theoretical objects and postulated interactions to get the desired intrinsic value, one has to admit that ones entity has no physical characteristic as such.

 

Further more:

In sum, the quark masses postulated according to the SM do not yield the nucleon masses by far. Gluons are massless.

Postulated Up-Quark mass: 2.3 ± 0.7 ± 0.5 MeV / c² up (u)

Postulated down-quark mass: 4.8 ± 0.5 ± 0.3 MeV / c² down (d)

938,272 0813 (58) MeV / c² Proton mass duu ~ 0,8 - 1,2% (!!!) Quark mass fraction

939,565 4133 (58) MeV / c² neutron mass ddu ~ 1,1 - 1,4% (!!!) Quark mass fraction

Thus, also heavy ions composed of protons and neutrons (such as lead or gold nuclei) can not be represented by quarks and gluons. This means that according to the principle of mass-energy equivalence, nucleons and, ultimately, heavy ions consist almost entirely of phenomenologically indeterminate binding energy. Even more complicated is the fact that the ions are accelerated to almost the speed of light before they collide. This means that there is also a considerable amount of external energy added to the binding energy. Neither the theory of relativity neither the SM does tell us how these phenomenologically can be divided into translational energy and "mass equivalence."

Protagonists of the SM are so convinced of their belief that they have obviously lost sight of the essential. Why should a postulated complex, multi-object-asymmetric, charge-fragmented, dynamic substructure create a spin value ½ and an elementary charge of exactly 1·e over dynamic states in the temporal or statistical mean? The comparison with the SM point-postulated, "leptonic" electron, with spin value ½ and elementary charge 1·e, which are "created" without "dynamic effort" and structure, identifies the quarks-gluon thesis as a fairy tale.  

 

The Quark Parton Model describes a complex theory situation and associated theory-laden (high-energy) experiments, the understanding of which is characterized by diverse aspects and a whole series of postulates. How the (here relatively simply described) nested theory-experiment-construct looks like, see representative the explanations (part 1 -3) of the Institute for Nuclear and Particle Physics Dresden.

Part1 fails early on the interpretations of the presented model based on "theory-internal" assumptions.

... "This is remarkable, especially the neutron as an electrically neutral particle should have vanishing magnetic moment. This already indicates that protons and neutrons are not point-like, but have an internal structure..."

Is the magnetic moment of the neutron really evidence for a substructure? Or is this assumption just a theory-laden measurement interpretation of the Standard Model?

The interpretation of a neutron-quark (sub)structure loses its meaning if one examines the situation numerically (independent of the thought model). The supposedly anomalous magnetic moments of electron and proton as well as of the neutron are finally a combination of the "semi-classical" - most simply calculated - "normal" magnetic moments and measurement-inherent contributions, which originate from the magnetic field, which is used for the measurement. This qualitative statement can be concretized (in accordance with the measured values). See further the remarks on the neutron from the point of view of a mass-radius-coupled model, with which both the mass and the magnetic moment of the neutron can be calculated from the neutron-forming interaction of electron and proton in a phenomenologically understandable and exact way.

The quark-parton model is an excellent teaching piece from the series of creeds (like once the epicycle or phlogiston theory). Experimental result interpretations led to some theoretical assumptions, which then quickly reach a substructured complexity by means of further interpretations with the help of new theory objects and postulated interactions, which finally eliminate any inappropriate experimental results by appropriate measures. First there were the postulated quarks. These required a fragmentation of the elementary electric charge. Then followed gluons, since quarks did not provide corresponding spin contributions of the nucleons. After that, so-called "sea quarks" were introduced as a supplement, since the postulated gluons also did not provide a corresponding spin contribution. Details of the "established" nucleon model see part 2 and part 3 (The Quark-Parton-Model) ... The analysis of the believers ends then with the belief finale

... "The nucleon is therefore built up from valence quarks, sea quarks and gluons. But for the properties of the nucleon (and all hadrons) like charge, mass and spin only the composition of the valence quarks is responsible. Therefore one often speaks of constituent quarks and means the valence quarks plus the surrounding cloud of sea quarks and gluons. Sea quarks and gluons do not contribute to the (net) quantum numbers, but they do contribute to the mass of the hadron, since they carry energy and momentum. While the bare mass of the valence quarks is only a few MeV/c², the constituent quarks carry the respective fraction of the hadron mass, in the case of the nucleon thus about 300 MeV/c². This is thus predominantly contributed by the gluons and sea quarks. In other words, the vast majority of the hadron masses (and thus the visible mass in the universe) is not carried by the bare masses of the constituents, but is dynamically generated by the energy of the interaction!" 

Isn't that nice. There are clouds and naked and the God-analogous indeterminate, which is about 99%, given the postulated masses of the "matter-forming" quarks. But here the faith fairy tale is not really at an end. In a bigger picture (ΛCDM-model) it is now postulated that the visible mass is again only a small part of the mass in the universe. For the most part, for today's Standard Model believers, the universe consists of Dark Matter and Dark Energy. Dark stands here for not detectable. No miracle that the Vatican invites year after year highly delighted to the science prayer. That neither the preachers nor the people a light rises has Theodor Fontane so formulated:

We are already deep in decadence. The sensational is valid and only one thing the crowd favours even more enthusiastically, the bare nonsense."

 

 

The "fragmentation of matter" as an »end in itself« of mathematical theories and the inevitable increase of irrelevant knowledge, especially in the form of virtual particles, has become established standard thinking. Instead of simplification, the concepts of formal postulations and "refining theories" obviously do not end in the growth of knowledge but in scientific arbitrariness. Mathematical-based fundamental physics urgently requires a natural-philosophical oriented regulation.

 

You might consider the following notes about neutrinos

There is no single direct neutrino proof. They are always strongly theory-laden interpretations of experimental results.

To generate the fermion masses by coupling the fermions to the Higgs field, the following conditions must be met: The masses of the right-handed and left-handed fermions must be equal. The neutrino must remain massless. This basic condition is in blatant contradiction to neutrino oscillations (Nobel Prize 2015), which require neutrino masses. Consequence: Either we say goodbye to the Standard Model of particle physics (SM) or to neutrinos with mass.

For the sake of completeness, it should be mentioned that ...since one does not want to part with the Standard Model, there is always the possibility of "somehow" incorporating neutrino masses within the framework of mathematical additions. However, this is only possible because the construct of the SM has no verifiable real-physics reference at all, i.e. all SM-associated alleged object proofs are indirect, strongly theory-laden experimental interpretations.

Energetically, no binding process of the transformation of a d-quark into a u-quark exists.

Independent of the SM-constructed lepton number, the following should be remembered: The neutrino was introduced historically because the energy spectrum of the (emitted) electrons does not show a discrete but a continuous distribution. But if the anti-electron neutrino, with whatever lower mass limit, "kidnaps" the "missing" energy from the laboratory system and can only "act" (effectively*) through the Weak Interaction, then this simply means that a kinetic energy continuum of the postulated neutrino already had to exist during the process of the Weak Interaction. Because after this process, according to the postulate, there is no further possibility of interaction. But how is this to be explained phenomenologically?

The terse statement of the SM that the beta-minus decay of the neutron takes place according to the conversion of a d-quark into a u-quark by means of a negatively charged W-boson says nothing about the concrete process of how, from where and why the anti-electron neutrino now absorbs different amounts of energy during the weak interaction in order to compensate for the "missing" energy in the electron spectrum.

On closer inspection, the situation is far more complex, since both postulated quark-based neutron and quark-based proton consist of ~ 99% of undefined binding energy and thus the weak interaction (energetically) affects only ~ 1% of the decay process.

The postulated mass of the d quark is 4.8 (+0.5 / - 0.3) MeV/c², the mass of the u quark is 2.3 (+0.7 / - 0.5) MeV/c²  quark masses : http://pdg.lbl.gov/2013/tables/rpp2013-sum-quarks.pdf

This means that the mass difference is between 1.5 and 3.5 MeV/c². The electron anti-neutrino with a mass (lower limit) of ≤ 2.2 eV/c² can absorb a maximum of ~ 0.78 MeV. According to the electron energy spectrum, however, the average energy is much smaller than 0.78 MeV, which is "abducted from the laboratory system" by the neutrino. What happened to the missing weak interaction energy? "Gluonic binding energy" it could not have become, since gluons do not participate in the postulated transformation of a d-quark into a U-quark. Nor can the virtual magic of the ~ 80.4 GeV/c² W boson absorb anything in real-energy terms. Here, it is more true than ever for the prevailing physics: "It is important to realize that in physics today, we have no knowledge of what energy is."...

So in the picture of the SM we have initially 1 u-quark and 2 d-quarks, in the meantime a W-boson and ~ 99% binding energy (whatever that should be phenomenologically) and after the transformation 2 u-quarks, 1d-quark, 99% binding energy, 1 anti-electron neutrino, 1 electron and additionally ~ 0.78 MeV energy. No matter how the energy distribution possibilities of the ~ 0.78 MeV to proton (2 u-quarks, 1d-quark, 99% binding energy), electron and anti-electron neutrino look like, the process of the weak interaction would already have to be such that these distribution possibilities are guaranteed, since after the weak interaction no more energy transfer of the neutrino to the proton and electron is possible. This means, however, that there can be no discrete transformation process of a d quark into a u quark.

Non-discrete means here in particular: Energetically, there is no binding process of the transformation of a d-quark into a u-quark. From a neutron-u quark, energetically different electron-anti-neutrinos are created by the weak interaction, since after the creation, other interaction possibilities apart from the weak interaction are excluded. Thus, there is no energetically unambiguous transformation of a u quark into a d quark.

Detached from this problem, it should also be remembered for the sake of completeness that experimentally we are not dealing with single objects but with many-particle objects (more than 1 neutron) and accelerated charges radiate energy. It is probably not to be assumed that the protons and electrons created "suddenly" exist with a constant speed. "Where" is the associated photon spectrum of beta-minus decay? What does it look like?

According to SM, neutrino research "means", for example:

...One measures the currents of the kaons and pions and indirectly determines the flux of the neutrinos...

However, pions (π π π+) and even more so kaons (K+, K-, K0, K0) are highly constructed entities (for initial understanding see https://en.wikipedia.org/wiki/Kaon , https://en.wikipedia.org/wiki/Pion ) of the Standard Model. Means: The number of existence postulates, such as mass, charge, spin, flavour(s), lifetimes and quark composition is already "considerable". The possible transformations result in "manifold alternation-interaction-fantasy scenarios”. Furthermore: The neutral kaon is not its own "antiparticle", this leads (more generally) to the construction of the particle-antiparticle oscillation (see https://en.wikipedia.org/wiki/Neutral_particle_oscillation ) and the neutral kaon is said to exist in two forms, a long-lived and a short-lived form.

To conclude from this that there are properties of "flavor-oscillating" neutrinos increases the arbitrariness even more. To understand all this (reproducibly), one needs absolute faith in axiomatic creations. As a reward, however, there is then the carte blanche that any experimental result becomes "explainable" ...that until then - in connection with the "experimental side" - we have to make, roughly estimated, several dozen further result-oriented assumptions, ... does not matter to SM-believers.

To quote Egbert Scheunemann: "They first shoot an arrow at a barn door, then paint concentric circles around the bullet and cheer about their "bull's eye".

Without question, ad hoc hypotheses and thought experiments were and are fundamental concepts. It is just that "physical" theses must sooner or later be substantiated by stringent, consistent “models of thought” using simple mathematical means, if they are to have any epistemological value. The bad habit of iteratively glossing over missing knowledge and missing phenomenology within the framework of parametric creeds etc. means arbitrariness and knowledge standstill. Renormalisation, variable coupling constants, free parameters, endless substructuring, perturbation theory with an affinity to Taylor series just correct the "physical crap" that one has primarily brought upon oneself over decades through a lack of insight.

By the way…At a conference in Rome in 1931, Niels Bohr expressed the view that understanding beta decay did not require new particles, but rather a similar serious overthrow of existing ideas as in quantum mechanics. He doubted the theorem of the conservation of energy, but without having developed a concrete counter-proposal…

 

 

Unfortunately there is no complete English translation for the "Elementarkörpertheorie" yet available. You'll find more detailed information if you select certain main issues from the website menu (auf Deutsch). "Feel free" to use a common webbrowser translation tool. You'll discover useful information, insights and surprising equations to deduce and calculate physical values based on mass-radius-relations such as... Sommerfeld Fine-structure constant, neutron mass, mass(es) of charged pions, mass and radius of the universe, Planck units, cosmic microwave background temperature, ...

 

                 [ Email-contact: Dirk Freyling : dfreyling@fastmail.net ]

 

... "for now" you'll find "here" important results and a short discription of how to gain those ...

 

Anatomy of anomalous magnetic moments

Based on the experimental values for the magnetic moments of electron, proton and neutron, it is concluded that the magnetic field itself provides an additive, measurement-inherent contribution ∆μB to the supposedly intrinsic values.

 

∆μB(electron)  ~  ∆μB(proton)  ~  ∆μB(neutron)    1 · 10-26 Joule/Tesla

 

Already for rational logical reasons, the question arises if these contributions (μe , ∆μp , ∆μn) have a common cause. The assumption presented here is that these contributions to magnetic moments are external magnetic field measurement-inherently contributions and are not intrinsic. This is investigated and confirmed numerically, analytically, phenomenologically as well as formally.

 

to be continued...