The Hierarchy Problem and the Cosmological Constant Problem Revisited

We argue that the Standard Model (SM) in the Higgs phase does not suffer from a “hierarchy problem” and that similarly the “cosmological constant problem” resolves itself if we understand the SM as a low energy effective theory emerging from a cutoff-medium at the Planck scale. We actually take serious Veltman’s “The Infrared–Ultraviolet Connection” addressing the issue of quadratic divergences and the related huge radiative correction predicted by the SM in the relationship between the bare and the renormalized theory, usually called “the hierarchy problem” and claimed that this has to be cured. We discuss these issues under the condition of a stable Higgs vacuum, which allows extending the SM up to the Planck cutoff. The bare Higgs boson mass then changes sign below the Planck scale, such that the SM in the early universe is in the symmetric phase. The cutoff enhanced Higgs mass term as well as the quartically enhanced cosmological constant term provide a large positive dark energy that triggers the inflation of the early universe. Reheating follows via the decays of the four unstable heavy Higgs particles, predominantly into top–antitop pairs, which at this stage are massless. Preheating is suppressed in SM inflation since in the symmetric phase bosonic decay channels are absent at tree level. The coefficients of the shift between bare and renormalized Higgs mass as well as of the shift between bare and renormalized vacuum energy density exhibit close-by zeros at about 7.7×1014GeV\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$7.7 \times 10^{14}\ \hbox {GeV}$$\end{document} and 3.1×1015GeV\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3.1 \times 10^{15}\ \hbox {GeV}$$\end{document}, respectively. The zero of the Higgs mass counter term triggers the electroweak phase transition, from the low energy Higgs phase and to the symmetric phase above the transition point. Since inflation tunes the total energy density to take the critical value of a flat universe and all contributing components are positive, it is obvious that the cosmological constant today is naturally a substantial fraction of the total critical density. Thus taking cutoff enhanced corrections seriously the Higgs system provides besides the masses of particles in the Higgs phase also dark energy, inflation and reheating in the early universe. The main unsolved problem in our context remains the origin of dark matter. Higgs inflation is possible and likely even unavoidable provided new physics does not disturb the known relevant SM properties substantially. The scenario highly favors to understand the SM and its main properties as a natural structure emerging at long distance. This in particular concerns the SM symmetry patterns and their consequences.

energy, inflation and reheating in the early universe. The main unsolved problem in our context remains the origin of dark matter. Higgs inflation is possible and likely even unavoidable provided new physics does not disturb the known relevant SM properties substantially. The scenario highly favors to understand the SM and its main properties as a natural structure emerging at long distance. This in particular concerns the SM symmetry patterns and their consequences.
Keywords Higgs vacuum stability · Hierarchy problem · Cosmological constant problem · Inflation

Prelude: Higgs Inflation in a Nutshell
In order to give a quick overview of what will be the essential conclusion of the analysis I start with this prelude (see [1,2]). The Standard Model (SM) hierarchy problem [3] is well known and addressed very frequently to motivate Beyond the Standard Model (BSM) scenarios in general and a supersymmetric extension of the SM in particular. The renormalized Higgs boson mass is small, at the ElectroWeak (EW) scale, the bare one is huge due to radiative corrections growing quadratically with the ultraviolet (UV) cutoff, which is assumed to be given by the Planck scale Λ Pl ∼ 10 19 GeV. 1 The cutoff dependence is illustrated in Fig. 1 assuming the cutoff as a renormalization scale and the SM Renormalization Group (RG) ruling the scale dependence of the SM couplings (see below). It is the RG improved version of Veltman's "The Infrared-Ultraviolet Connection" [4], where the SM renormalization of the Higgs boson mass (m 0 the bare, m the renormalized mass) has been addressed (see also [5][6][7][8]). The coefficient function C(μ) depends on the dimensionless SM couplings, which depend on the renormalization scale logarithmically only. For an early discussion of the impact of running couplings to the Higgs 1 The Planck medium, which we may call ether, somehow gets shaped by gravity and quasi-particle interactions emergent in the SM at low energies. It is characterized by the well known fundamental Planck cutoff Λ Pl or equivalently the Planck mass M Pl , which derive from the basic fundamental constants, the speed of light c characterizing special relativity, the Planck constant intrinsic to quantum physics and Newton's constant G N the dimensionful key parameter of gravity. Unified they provide an intrinsic length Pl , the Planck length, which also translates into the Planck time t Pl and the Planck temperature T Pl : Planck length: Pl = In our context they define a shortest distance Pl and a beginning of time t Pl . i.e. t > t Pl . The Planck era energy scale equivalently is set by E Pl = Λ Pl ≡ M Pl or temperature T Pl , as for most time in the evolution of the early universe, when elementary particle physics is at work and before the epoch of formation of hadrons, particle processes are in thermal equilibrium, with well known exceptions during inflation and the electroweak phase transition. Pl . The effective mass is given by the bare mass at short distances and by the renormalized one at low energy mass term and the problem of fine-tuning see [9]. The Higgs mass counter term is huge when we adopt the Planck scale as the cutoff to regulate UV singularities. Is this a problem? Is this unnatural? In the first instance it is a prediction of the SM! At low energy we see what we see (what is to be seen): the renormalizable, renormalized SM [10][11][12][13] as it describes close to all we know up to LHC energies. But what does the SM look like if we go to very high energies even to the Planck scale? Not too far below the Planck scale we start to see the bare theory i.e. the SM with its bare short distance effective parameters, so in particular a very heavy Higgs boson, which likely is moving at most very slowly. The potential energy then is large while the kinetic energy 1 2φ 2 is small, as a dedicated calculation shows. Here we have in mind the cosmological solutions of Einstein's General Relativity Theory (GRT) for an isotropic universe of constant spatial curvature, parametrized by the Robertson-Walker metric. 2 The field φ is then a function of the cosmic time t only andφ is the corresponding time derivative. The Higgs boson contributes to the 2 Einstein's field equation for the metric tensor g μν , which incorporates the gravitational field, is given by 2 is the effective interaction constant, G μν = R μν − 1 2 R g μν is the Einstein curvature tensor (geometry) and T μν is the energy-momentum tensor (matter and radiation). Cosmology is shaped by Einstein gravity, which together with Weyl's postulate, that radiation and matter (galaxies etc.) on the cosmological scale behave like an ideal fluid, and the cosmological principle, assuming isotropy of space (implying homogeneity), fixes the form of the metric and of the energy-momentum tensor: (1) the metric (3-spaces of constant curvature k = ±1, 0) takes the form ds 2 = (cdt) 2 −a 2 (t) [dr 2 /(1 − kr 2 )+r 2 dΩ 2 ], where in the comoving frame ds = c dt with t the cosmic time; (2) the energy-momentum tensor takes the form T μν = (ρ(t) + p(t)) (t) u μ u ν − p(t) g μν ; u μ .
= dx μ ds where ρ(t) is the density and p(t) the pressure of the fluid. As a differential equation for the geometry factor a(t) one obtains Friedmann's equations (3). One needs ρ(t) and p(t) (which are related by an equation of state characterizing the medium) in order to get the radius of the universe a(t) and its evolution in time. The Higgs potential contributes T μν = Θ μν = V (φ) g μν + derivative terms, where Θ μν is the symmetric energy momentum tensor of the energy momentum tensor by providing the pressure p = 1 2φ 2 − V (φ) and the energy density ρ = 1 2φ 2 + V (φ) . As we approach the Planck scale (bare theory) the slowroll condition 1 2φ 2 V (φ) is satisfied during some window in the time evolution and then p ≈ −V (φ) ; ρ ≈ +V (φ) implies p ≈ −ρ, which closely approximates the equation of state p Λ = −ρ Λ of Dark Energy (DE) ρ = ρ Λ , the equivalent of a Cosmological Constant (CC) Λ. DE follows a very special equation of state, only observed indirectly through Cosmic Microwave Background (CMB) [14][15][16][17][18] pattern and through Super Novae (SN) counts [19,20]. No lab system observation so far has been reported to my knowledge, although statistical mechanics system like the Ising model obey such a ground state equation (see e.g. [21]). Thus, as a consequence of the hierarchy boost, the SM Higgs boson in the early universe delivers a huge dark energy that is inflating the universe and, which mimics strong anti-gravity at work. The Friedmann equations together with energy conservation read and indeed if DE dominates we have (3 p+ρ) ≈ −2ρ such that we have an accelerated expansionä/a > 0 and da a = H (t) dt which implies an exponential growth a(t) = exp Ht of the radius a(t) of the universe. H (t) =ȧ(t)/a(t) is the Hubble constant where H ∝ √ V (φ) in a DE dominated era. Inflation stops quite quickly as the field decays exponentially. The field equation for a potential dominated by the mass term V (φ) ≈ m 2 2 φ 2 represents a harmonic oscillator with friction and leads to Gaussian inflation as established by an analysis of the CMB pattern by the Planck mission [22]. One of the reasons why the inflation phenomenon must have happened in the early universe is that the universe looks flat today, while a flat universe in the absence of DE is exponentially unstable in its time evolution. This because then formally the strong energy condition ρ + 3 p ≥ 0 holds, which impliesä/a ≤ 0. So, different types of solutions of the Friedmann equations at ρ Λ = 0 deviate dramatically during the 13.8 billion years life of the universe after the Big Bang. The "flattenization" by inflation is evident as the curvature term k/a 2 (t) ∼ k exp(−2Ht) → 0 drops exponentially independent of the curvature type. The latter, characterized by the normalized curvature k = 0, ±1, distinguishes flat infinite, spherical closed or hyperbolic open geometries. It is very important to note that the CC given by the Higgs potential V (φ) in the symmetric phase is positive in any case, very different from the (much smaller) contribution from the Higgs-field vacuum expectation value (VEV) in the broken phase, which is negative [23]. This already shows that dynamical effects and in particular possible phase transitions make the CC depend on time, although it does not depend explicitly on a(t), i.e., ρ Λ ∝ a(t) 0 . As we Footnote 2 continued SM (or extensions of it). Only a scalar potential can contribute a term proportional to g μν , which mimics a cosmological constant. know the matter density scales like ρ m ∝ a(t) −3 while the radiation density decreases like ρ γ ∝ a(t) −4 since radiation gets red-shifted in addition to its spatial dilution. 3 Inflation tunes the total energy density to be that of a flat space (as if k = 0), which according to (3) for k = 0 requires a specific "critical" energy density ρ 0,crit = 3H 0 /(8π G N ) = μ 4 crit where μ crit ≈ 0.00247 eV .
With H 0 the present Hubble constant ρ 0,crit is the present total energy density. For an arbitrary mixture of dark energy, matter and radiation the first of the Eq. (3) reads where Ω i = ρ 0,i /ρ 0,crit are the present fractional densities and a 0 = a(t 0 ) the spatial metric scale factor at present time t 0 . Including the curvature term Ω k = −k/(a 2 0 H 2 0 ) we have as an exact equation, and when the curvature term is exponentially suppressed we very accurately have what is supported strongly by observation (CMB). Whatever constitutes the universe, the curvature constant is k = +1, k = 0 or k = −1 according to whether the present density ρ 0 is greater than, equal to, or less than ρ 0,crit . A higher density ρ 0 > ρ 0,crit implies a recontraction of the initially (at the Big Bang) expanding universe, a lower density ρ 0 < ρ 0,crit would not be able to stop the expansion forever. We know that ρ 0Λ = μ 4 0Λ today is about μ 0Λ = 0.00171 eV which in a flat universe must be a fraction of the critical density, and actually has been determined to amount to 69.2 ± 1.2 %. Since the non-DE components drop with a power of the radius a(t) as time goes on, the asymptotic behavior is determined by ρ Λ solely. Friedmann solutions in GRT with non-vanishing cosmological constant have been discussed in [24,25].
We note that the large positive cosmological constant provided by the SM Higgs sector on the one hand effectuates the inflation that is needed to tune the total energy density into the critical density of flat space. This tells us that on the other hand the cosmological constant has to be some fraction of the critical density, i.e. it is self-tuned to be small. So also the cosmological constant problem may turn out to get its natural explanation (see below).
Since inflation is strongly supported to have happened by observation, we must assume the existence of an appropriate scalar field, and the Higgs field is precisely such a field we need and within the SM it has the properties that promote it to be the inflaton. In contrast to other inflaton models the Higgs inflaton is special because almost all properties are known or predictable! Below, I will argue that the SM in the Higgs phase does not suffer from a "hierarchy problem" and that similarly the "cosmological constant problem" resolves itself if we understand the SM as a low energy effective theory emerging from a cutoff-medium at the Planck scale.
Adopting a bottom-up approach, I discuss these issues under the condition of a stable Higgs vacuum, by predicting the behavior of the SM when approaching the Planck era at high energies. SM Higgs inflation as exposed in this prelude may look pretty simple but in fact is rather subtle, because of the high sensitivity to the SM parameters and a high sensitivity to higher order SM effects. In any case my preconditions are: (i) a stable Higgs vacuum and a sufficiently large Higgs field at M Pl , (ii) physics beyond the SM should not spoil main features of the SM. This means that SM extensions like SuperSymmetry (SUSY), or Grand Unified Theories (GUT) etc., pretending to solve the hierarchy problem and/or affecting the SM RG-flow substantially, are to be excluded! Here we have to assume that a kind of desert in the heavy particle spectrum is extending effectively up to the Planck scale. This is not so far beyond the "grand desert" usually assumed to exist in the context of GUTs. This does not exclude new physics that we know to exist, like dark matter, Majorana neutrinos or axions, for example.
In this prelude I have outlined what a correct interpretation of the "hierarchy problem" likely looks like, i.e. the predicted SM hierarchy pattern is not a problem, rather it is the solution for what we need to trigger inflation in the early universe. In the following I will consider the hierarchy issue in a broader context and discuss in some detail the intricacies of the cosmological constant problem and Higgs boson inflation. I will try to convince the reader that the Higgs boson inevitably delivers dark energy and the consequent inflation is well supported by a self-consistent perturbative SM calculation [1,2]. The approach is highly predictive and limited mainly by the uncertainties of the knowledge of the SM parameters and the accuracy of the perturbative calculations of the matching conditions between measured and MS parameters and the MS renormalization group coefficients.

The Higgs Boson Discovery: The SM Completion
With the discovery of the Higgs boson by ATLAS [46] and CMS [47] in 2012 a last major but often questioned building block of the electroweak SM has been experimentally verified. The existence of an elementary scalar has been found to be required   [65] to render the electroweak massive gauge theory renormalizable in 1964 by Englert and Brout [48] and by Higgs [49]. The key mechanism turned out to be a Spontaneous Symmetry Breaking (SSB) mechanism of the non-Abelian SU(2) L gauge sector responsible for the weak interactions. The corresponding Higgs mechanism generates masses for all massive particles while not affecting the renormalizable UV behavior of the massless unbroken theory. Now, remarkably, the SM Higgs boson mass has been found in very special mass range 125.18 ± 0.16 GeV, which seems to match the possibility to extrapolate the SM up to the Planck scale. Knowing the Higgs mass M H and using the mass coupling relation valid in the Higgs phase, we also know the Higgs self-coupling λ and hence the renormalized Higgs potential V = m 2 2 H 2 + λ 24 H 4 , which is the object in our focus. Perturbativeness and vacuum stability of the Higgs potential are the key issues in this context (for early considerations see [50][51][52][53][54][55]). Figure 2 adopted from an analysis by Hambye and Riesselmann in 1996 [56] illustrates the possible impact to have a Higgs mass in a window extending up to the Planck scale. The possibility that the SM may be extended right up to the Planck scale has been analyzed also in [57][58][59][60][61][62][63][64]. Later estimates have been improved after more precise SM parameters like the QCD coupling α s and top-quark mass M t became available, see e.g. [65][66][67][68][69]. Given the Higgs self-coupling, all relevant SM parameters are known. While the RG evolution equations in the symmetric phase of the SM have been known for a long time to two loops, recently also the three-loop coefficients in the MS scheme have been calculated [70][71][72][73][74][75][76][77]. The MS input parameters, which are most suitable to parametrize the high energy tail, have to be calculated via the matching conditions from the experimentally measured ones (see [65,68,69,[78][79][80][81][82][83][84][85][86] and references therein). The matching conditions are based on more or less complete two-loop calculations that require the knowledge of the two-loop renormalization counterterms. Besides the electromagnetic vertex correction δe, all others are mass renormalization counter terms δ M 2 b for bosons and δ M f for fermions, all given in terms of comparatively simple two point functions, which are completely known to two loops. 4 So far so good. One important point is missing however: the physical on-shell parameters are determined from experimental data by unfolding the raw data from radiation and detector effects. They represent pseudo-observables depending on theory input that relies on approximations. One has to keep in mind that most LEP, Tevatron and LHC results, are based on incomplete two-loop calculations at best. Exceptions are the extraction of G μ [92] and of sin 2 Θ lept eff for which complete two-loop calculations exist [93][94][95]. Complete two-loop calculations have not been available at LEP times. For Bhabha scattering, which plays a key role for luminosity monitoring, the two-loop QED corrections became available only lately (see e.g. [96] and references therein). Full two-loop electroweak corrections to 2 → 2 processes either are still missing or have not been available when parameters were extracted from the data. There are persisting discrepancies at the 2 σ level in the determination of sin 2 Θ lept eff between SLD and LEP as well as for sin 2 Θ bb eff and so there remain questions about the size of the estimated uncertainties of some of the input parameters.
Precision physics at LEP has been a great achievement also thanks to continuous progress in QCD and electroweak higher order calculations. Much progress has been achieved since. The SM is established with unprecedented accuracy. Activities now are focusing on LHC physics for obvious reasons. But more effort is needed to keep alive gained expertise on electroweak precision physics. Projects for future e + e − colliders like the International Linear Collider (ILC), which started with the unrealized TESLA project [97], and the Future Circular Collider (FCC) project at CERN, are boosting efforts to reach much higher precision in Z -peak, W -pair, top-quark and Higgs-boson physics [98]. The possibility that progress in this direction can establish the Higgs vacuum to be stable and the Higgs boson to be the inflaton in fact may be the strongest motivation to go on in realizing a next e + e − machine probing physics at the electroweak scale at much higher precision.
The most serious conceptual problem concerns the measurement of the top-quark mass and the related determination of the top Yukawa coupling. On the theory side one is operating with the perturbative concept of an on-shell top-quark mass, which is not an observable. Usually, the on-shell mass is identified with the Monte Carlo mass M MC t . The latter is the kinematic mass parameter used Monte Carlo event generators that are utilized to extract the top-quark mass from top-quark production processes. The non-perturbative color screening effects obscure the precise determination of M t so far (see e.g. [69,99]). In addition, most of the Monte Carlo programs used at present do not take into account the electroweak radiative corrections.
My quintessence: while the matching conditions are known to two loops, the input parameters have not yet been determined at the same level of accuracy, i.e. likely 4 A complete two-loop calculation has been performed independently by Veretin and Kalmykov in the context of [80][81][82]. Two-loop integrals exhibiting Higgs propagators have been expanded in M 2 V /M 2 H , assuming the Higgs to be heavier than the W and Z bosons. After the Higgs mass has been found this expansion turned out to be obsolete, and the relevant integrals had to be evaluated numerically. This has later been performed partially in [83,84] and "exact" in [87] such that a complete two-loop evaluation is available. For independent calculations of the matching conditions also see [88][89][90][91] and references therein.
reported errors are underestimated. These issues have to be reminded before one can claim that metastability of the electroweak vacuum is a proven fact.

Matching Conditions: MS Parameters in Terms of Physical Parameters
We want to solve the RG equations up to very high scales, where mass effects are supposed to be negligible, in regions where physical thresholds play no role. The mass independent MS scheme is the most simple choice, and under these circumstances likely a rather physical one, because it reflects the true UV structure of the SM in terms of quasi-bare quantities. In order to solve the MS RG-equations we need the input MS values of the basic couplings g , g, g 3 , y t and λ. While is a standard MS reference parameter, the key parameters y t and λ are known actually only through the related masses M t and M H : where δ i represent the corresponding radiative corrections. The relations between the renormalized on-shell masses and their MS versions are provided by the natural matching conditions for bosons and m f ren = M f ren + δ M f − δm f , for fermions. 5 Formally we obtain them by writing the relation between the renormalized and the bare masses (in the bare Lagrangian) in the two schemes: for bosons and for fermions, i.e. the mass shift is given by the MS finite part prescription applied to the (UV singular) OS mass counterterm: for bosons and for fermions.
Here it is important to keep in mind: for the renormalized theory the MS parametrization is not a parametrization in terms of observables but serves as a convenient intermediate (auxiliary) parametrization. The MS scheme is a purely perturbative Tadpoles are gauge dependent and UV singular, and have to be taken into account as shown in order to preserve the gauge symmetry [78] concept (no non-perturbative definition) and corresponding parameters are not measurable. However, if we take the bare theory to be the physical one in the sense of a low energy effective theory exhibiting a physical cutoff, then the MS parameters in the perturbative regime are representing the bare parameters. Crucial for the extrapolation to the Planck scale, if possible, it is to keep the relationships between bare, MS and physical OS parameters gauge invariant and preserving the UV structure (see e.g. [78,79]).
So in principle, on a conceptual level, we are confronted with a well-defined problem of calculating the massive physical particle self-energies exact to two loops, in addition to the e.m. vertex, in order to find the appropriate input for the three-loop MS RG running.
In the symmetric phase of the SM, except from the m 2 -term in the Higgs potential, all masses vanish and the RG coefficient calculations and the solution of the RG equations are straight forward. This insofar that there are no special mass-coupling relations and the decoupling of heavy states (the four Higgs scalars) is given as requested by the Appelquist-Carazzone theorem (AC) [100], which in the broken phase only holds in the QCD and QED sectors.
In contrast, in the Higgs phase of the SM there are some tricky points to be taken care of.
(1) The tadpole issue: if we require the bare parameters, now represented by the MS parameters, as physical we have to respect Ward-Takahashi and Slavnov-Taylor identities. This requires to take into account tadpole contributions according to Fig. 3, which mostly are omitted in calculations (see e.g. [86,[88][89][90][91]101]), because a theorem [102,103] states that tadpoles cancel in physical quantities provided they are expressed in terms of physical quantities within the renormalized theory (see [83,84] for a recent discussion). (2) The lack of decoupling issue: while in QED and QCD, heavy particles decouple, within the SM heavy states do not decouple when the mass-coupling relations come into play. Masses and couplings are one-to-one interrelated, because all masses are generated by the Higgs mechanism. For the given VEV v a mass can only get large iff the corresponding coupling gets large. The couplings are active at scales below the related mass thresholds. Within the MS scheme, which respects gauge invariance but has deficiencies like the lack of decoupling of heavy states, decoupling has to be imposed by hand, and one is working with effective field theory including only the active flavors at the given scale, in place of full theory. Most of the matching analyses are inspired by techniques that are well established in QCD. Typically, one is matching the N f with the N f + 1 flavor effective QCD at the N f + 1 flavor threshold. This is well justified in the time-like regime where corresponding thresholds are manifest. While physical observables are naturally sharing decoupling properties, within the full SM the lack of automatic decoupling is a serious short-coming of the MS parametrization when applied below the highest SM threshold at 2 M t .
The most prominent non-decoupling effect is due to the large top Yukawa coupling, which we know to be interrelated with the heavy top-quark mass. It yields the leading correction of the EW ρ-parameter, defined by the neutral to charged current ratio of the corresponding low energy effective Fermi couplings, which is given by where N c = 3 is the number of colors. It is quadratic in y t and measures weak-isospin breaking by the Yukawa couplings of the heavy fermions at zero momentum. So, the top Yukawa coupling is at work down to zero momentum and not being active starting above the top mass threshold only. This type of effect measured for the first time at LEP in 1995 far below the top-pair threshold, allowed to derive a bound on M t before the top-quark discovery at the Tevatron in 1996. This shows that "decoupling by hand" cannot be applied for weak contributions, i.e., we cannot parametrize and match together effective theories by switching off fields of mass M > μ at a given scale μ.
Only a direct measurement of y t and λ at a facility like FCC-ee or ILC above the top-quark mass threshold can provide us the precise input parameters we need.
(3) The Fermi constant issue: the Higgs-field VEV v determines the Fermi constant via For the on-shell counterterm we then have the relation (c 2 Potentially, the Higgs-field VEV v could be particularly affected by nondecoupling effects. However, here we may take advantage of the fact that tadpole contributions drop out in relations between physical (on-shell) parameters and physical transition amplitudes. To be compared are given by leptonic W decay rate. IndeedĜ μ ≈ G μ with good accuracy as expected. 6 Finally we need the MS version of G F . Again, evaluating G F bare = G μ + δG F in both schemes we have We calculate it equivalently by using (13) in the respective schemes. Then the MS top-quark Yukawa coupling is given by A good matching scale is M Z using α(M Z ), α s (M Z ) andĜ μ as input. The nonperturbative contribution Δα (5) had (M 2 Z ) to the shift in α(M Z ) is taken from [104]. I have outlined the points that can lead to slightly different input values for the MS parameters as listed in Table 1. Our evaluation is documented in more detail in Refs. [1,2,83,84]. One difference concerns the use of the Fermi constant, which in a number of analyses is taken to be the low energy G μ , assumed not to be running below the top-quark mass threshold, and often a difference between its OS and MS version is not made (see e.g. [65,85]). A second cause for a difference is related to a different account of the tadpole contributions (see e.g. [86,[88][89][90][91]). Given the MS input, the RG-equations are then solved in the MS scheme to three loops by also including four-and five-loop results in the QCD sector. Concerning the RG-running, there is full agreement between the different studies available in the literature for given set of input parameters. The only parameter which does not agree within quoted uncertainties is the top-quark Yukawa coupling y t (9), which apparently is the parameter which decides about the stability of the vacuum.
Another important point, which is only partially taken care of in estimating the uncertainties of the physical input parameters, is the following: (4) The scheme dependence issue: perturbative predictions are renormalization scheme dependent due to truncation errors of the perturbative expansion. A typical example is provided by the electromagnetic fine structure constant. At very low energy the value in the Thomson limit α 1/137 is adopted as an input parameter, while at the Z boson mass scale α = α(M Z ) = α 1−Δα(M Z ) 1/127 is a more appropriate input parameter, because it represents an universal type large correction requiring RG resummation, and which enters radiative correction calculations at many places. To one-loop order α = α (1 + a α). If we calculate a matrix-element including one-loop corrections in terms of α we obtain and hence Thus the result differs by δ M. If we do not actually calculate the higher orders (1) must be considered as an uncertainty due to unknown higher order effects. Actually, results can differ non-negligibly when different parameter sets are used as independent input parameters, like α, M W , M Z , . . ., G μ , M W , M Z , . . . or the preferred LEP parametrization in terms of the most precisely known parameters α, G μ , M Z , . . ., the scheme which is usually applied (see e.g. [105]).
I am mentioning the scheme dependence uncertainties here because to my knowledge they are not fully taken into account in standard extractions of on-shell parameters from the data. One more reason why strong statements concerning a proof of metastability and "just failing vacuum stability", it seems to me can hardly be justified in view of the strong sensitivity of vacuum stability on precise input parameters.
I remind that the most important difference between my "cutoff extended" SM and the most often discussed metastability path to high energies (discussed within the framework of the renormalized SM) lies in the fact the I have to take care that my MS parametrization is equivalent to the bare parametrization. This is what makes the inclusion of tadpoles mandatory. We have explicitly checked in [80][81][82] that only by including tadpoles the MS RG equations in the broken phase agree with the ones in the symmetric phase. 7  Table 1

The SM Running Parameters
In Fig. 4 we plot the evolution of the SM couplings as a function of the log of the energy scale.
As we learn from Fig. 4 the amazing thing is that the perturbation expansion turns out to work up to the Planck scale! In our analysis, for the input parameters specified below, we have no Landau pole or other singularities and the Higgs potential remains stable. This likely opens a new gate to precision cosmology of the early universe [1,65,68]. The remarkable interrelations between SM couplings may be summarized as follows: the U (1) Y coupling g 1 is screening (IR free), the SU(2) L coupling g 2 and the SU(3) c coupling g 3 are anti-screening (UV free=Asymptotic Freedom (AF)) as expected [106,107]. In contrast the top-quark Yukawa coupling y t and the Higgs self-coupling λ, which are screening if standalone (IR free, like QED), within the SM change their behavior from IR free to UV free, such that perturbation theory works the better the higher the energy in these couplings as well. What happens is that QCD effects dominate the behavior of the top-quark Yukawa coupling RG provided g 3 > 3 4 y t in the gaugeless (g 1 , g 2 = 0) approximation, which is satisfied. Similarly, the top-quark Yukawa effect dominates the Higgs coupling RG provided λ < 3 2 ( which also holds in the gaugeless (g 1 , g 2 = 0) limit. These conditions are satisfied in the SM with the given parameters and extend to higher orders as far as these are known. We note that the Abelian hypercharge coupling g 1 , although increasing with energy, stays small up to Λ Pl such that it does not affect perturbativeness. Note that in spite of its increasing behavior g 1 < g 3 < g 2 at Planck scale. Interestingly there we have an inverted g 3 < g 2 hierarchy of the non-Abelian gauge couplings. In the focus is the Higgs self-coupling, because it may not stay positive λ > 0 up to Λ Pl . In fact a 3 σ significance for meta-stability is claimed e.g. in [65,68] (see right panel of Fig. 2). Calculating previously missing two-loop contributions to the matching conditions the significance for missing stability could be reduced to a 1 σ gap in [86]. The existence of a zero in λ(μ) crucially depends on the precise size of the top-quark Yukawa coupling y t , which actually seems to decide about the stability of our world. Note that λ = 0 would be an essential singularity! Uncertainties here have to be reduced by more precise input parameters and better established EW matching conditions. For our input parameters the Higgs coupling decreases up to the zero of the beta-function β λ at μ λ ∼ 3.5 × 10 17 GeV, where λ is small but still positive and above μ λ increases with energy up to μ = M Pl .
I think our discussion shows that ATLAS and CMS results may have revolutionized particle physics in an unexpected way, namely showing that the SM has higher self-consistency (conspiracy) than expected and previous arguments for the existence of new physics may turn out not to be compelling. Also the absence so far of any established new physics signal at the LHC may indicate that commonly accepted expectations may not be satisfied. On the one hand it seems to look completely implausible to assume the SM to be essentially valid up to Planck energies, on the other hand the flood of speculations about physics beyond the SM have been of no avail. Within the context of GUTs, a large gap in the particle spectrum, the "grand desert" up to the GUT scale at about 10 16 GeV, still is a widely accepted hypothesis. So why the "grand desert" could not extend a little further namely to 10 19 GeV? The central issue for the future is the very delicate "acting together" between SM couplings, which makes the precision determination of SM parameters more important than ever. This mainly challenges accelerator physics, the LHC experiments and the future ILC and FCCee projects (top-quark and Higgs-boson factories), which could improve the precision values for λ, y t and α s . Still important are lower energy hadron facilities, which should provide more precise hadronic cross sections in order to reduce hadronic uncertainties in α(M Z ) and α 2 (M Z ). This could open a new gate to precision cosmology of the early universe, in case the Higgs boson inflation scenario outlined in Sect. 1 could be hardened.

Thoughts on Guiding Principles and Paradigms in Particle Physics
In last decades "solving the hierarchy problem" has been a strong motivation to find possible extensions of the SM. Guiding principles often have played an important role for progress in science although they afterwards turned out to miss the point they suggested natural laws should follow. Most prominent are symmetry principles. Related group theory is beautiful mathematics but is not always mapping the real or supposed physical problem it was proposed to describe. Kepler already dreamed of the Platonic bodies (regular polyhedra) to rule celestial mechanics of planets. After his attempt to prove this by analyzing celestial data, finally Kepler's laws resulted from his investigation. Kepler's model is completely false, the interplanetary distances it predicts are not sufficiently accurate, and Kepler was scientist enough to accept this eventually. But it is an excellent example of how truth and beauty do not always fit together. The widespread string theory paradigm assumes that a simple highly symmetric stringy structure at and beyond the Planck scale 8 could explain what we observe down on earth, actually a rather complex real world. That something simple looks complicated when seen from far away is certainly not a very natural expectation. A solid macroscopical may look to be perfectly rotational symmetric, zooming in on its microscopic structure can uncover a lattice of atoms exhibiting all kinds of lattice defects and domain patterns. The "the closer you look the more there is to see" pattern of thought looks to me less artificial than expecting some harmonic oscillator "heaven". In any case concepts like the ones we are discussing here: Naturalness, Hierarchy and Fine-Tuning make no sense without specifying the context in which they are addressed.
While "solving the hierarchy problem" seems to fail as a route to new physics, in contrast, the concept of the minimal renormalizable extension of Fermi's weak interaction theory turned out to be impressively successful in constructing piece by piece the electroweak SM. It has lead to the introduction of the massive intermediate spin 1 bosons W ± in charged current processes, the prediction of neural currents and the need for a Z boson. The resulting U (1) Y ⊗SU(2) L gauge theory, renormalizable in the massless case, required a scalar spin 0 boson, the Higgs boson, as a trick to generate masses of the weak gauge bosons and the fermions, without spoiling renormalizability. Spontaneous symmetry breaking, a mechanism known from condensed matter physics, turned out to be the key mechanism for a renormalizable massive gauge theory. Another example concerning minimality versus non-minimality we have when considering GUTs, where fermions are necessarily populating higher representations while the fundamental ones are not occupied. Therefore the typical leptoquarks necessarily showing up in GUTs are unnatural as an emergent phenomenon. 9 A convincing solution of the SM's hierarchy problem is known to be provided by a supersymmetric extension of the SM. SUSY cannot be an exact symmetry because it would predict a degenerate mass spectrum while phenomenologically the states of the SUSY mirror world all must be heavier than the SM particles. This leads to a very complicated world as a broken SUSY scenario not only is doubling the spectrum at once but also leaving too much freedom, with about 100 unknown symmetry breaking parameters. This makes such extensions not really predictive without additional assumptions. At the end phenomenological constraints require a SUSY version that would not be solving the hierarchy problem really, rather it would only be shifting the amount of fine-tuning required.
In addition, the hierarchy fine-tuning problem if being solved by a supersymmetrization of the SM creates new problems as we know. First, a second Higgs doublet field needs to be introduced, which as such is an interesting option. However, in order not to be in conflict with the absence of tree level Flavor Changing Neutral Currents (FCNC) 10 one has to impose R-parity, which is not less a fine tuning, although FCNCs can be forbidden by a simple discrete symmetry. R-parity is not required by 9 The unification paradigm celebrated its triumphal success in Maxwell's electromagnetism, which unified electrical and magnetic laws and predicted electromagnetic waves. In contrast as we know the electroweak theory is not a true unification, it rather regulates the mixing of electromagnetic and weak interaction phenomena. At the heart is γ − Z mixing and Z resonance (a kind of "heavy-light") physics, which manifests itself most convincingly in electron positron annihilation into Z bosons. All further unification attempts so far are missing confirmation. 10 FCNCs are automatically absent in the SM by the GIM mechanism [109] , as it is also highly established by experiment. renormalizability, it is not naturally emergent in a low energy effective theory and thus looks to be ad hoc. 11 A generic SUSY extension as such would be in contradiction with observation right away. This also illustrates that naturalness is a doubtful concept:

R-parity is a symmetry which forbids FCNCs but what is natural about it?
In my opinion the dogma surrounding the so-called hierarchy or fine tuning problem turned out to be a complete failure. Similarly, the GUT paradigm has not lead to any experimentally confirmed predictions that would support this concept. I think the minimal renormalizable QFT paradigm is back. Thereby "minimal" is crucial, many higher renormalizable structures like a GUT extension of the SM are not natural in that sense. Interestingly, a missing third fermion family or an additional fourth family would spoil important properties of the SM, such as the RG flow, and the Higgs boson then could not be a candidate for the inflaton. Another very important special feature of the SM is its tree level accidental custodial symmetry. The latter is violated by many of the proposed extensions of the SM, which then create a different fine tuning problem [110], in all cases that violate the tree level SM relation cos 2 This relation is strongly supported by experimental data, which precisely confirm SM predictions of the radiative corrections.
One also has to keep in mind that precision tests of the SM already revealed a test in-depth of its quantum structure. The largest corrections come from the running fine structure constant α(s), the running of the strong coupling α s (s) and the large top Yukawa y 2 t (s) effects. As contribution to the ρ = G NC /G μ (0) parameter, for example, subleading corrections amount to a 10 σ deviation from the SM prediction when taking into account the largest corrections only. Thus the SM is on very solid grounds better than everything else we ever had.
On the other hand the view that the SM is a low energy effective theory of some cutoff system at the Planck energy scale M Pl appears to be consolidated. This also puts QFT on a firm mathematical basis. A crucial point is that M Pl , providing the scale for the low energy expansion in powers E/M Pl , is exceedingly large, very far from what we can see! A dimension 6 operator at LHC energies is suppressed by (E LHC /Λ Pl ) 2 ≈ 10 −30 . This seems to motivate a change in paradigm from the view that the world looks simpler the higher the energy to a more natural scenario which understands the "cutoff SM" as the "true world" seen from far away, with symmetries emerging from not resolving the details of the short distance structure. In the low energy expansion one is "throwing away" an infinite tower of shorter distance information carried by the suppressed so called irrelevant operators.
The hierarchy problem requires to take the relationship between the bare UV and the renormalized IR regime as testable physics. Here Kenneth Wilson's 1971 solution [111,112] of a problem which has been persisting for about 75 years, surrounding the critical indices of phase transitions in condensed matter systems, has shed new light on the role cutoffs may play in physical laws. Wilson's renormalization semi-group, based on integrating out irrelevant details of the short distance structure, opened the quantitative approach of constructing low energy effective quantum field theories that derive from systems whose short distance structure have an intrinsic cutoff, like an atomic lattice or an atomic gas or fluid (see e.g. [113]). The key low energy emergent structure notably turned out to reveal renormalizable Euclidean quantum field theory. The latter exhibits analyticity in a way which makes it equivalent to a Minkowski quantum field theory. The latter hence is incorporating quantum mechanics as an emergent structure. As I will argue in the following, cutoffs in particle physics are unavoidable in understanding the relationship between a bare and a renormalized theory (see e.g. [114]). In such context renormalizability is an emergent property like all structures required in order renormalizability to be manifest. In our context cutoffs are indispensable for understanding early cosmology in a bottom-up way [1]. This opens the possibility of an alternative understanding of inflation, reheating, baryogenesis and all that [26][27][28][29][30][31][32]. As in condensed matter physics the connection between macroscopic long distance physics (at laboratory scales) and the microscopic underlying cutoff system (high energy events as they were natural in the early universe) turn out to have a physical meaning.
I remind that the SM's naturalness problems and fine-tuning problems have been made conscious by 't Hooft [3] long time ago as a possible problem in the relationship between macroscopic phenomena that follow from microscopic laws (a condensed matter system inspired scenario). Soon later the "hierarchy problem" had been dogmatized as a kind of fundamental principle. In fact the hierarchy problem of the SM had been the key motivation for a number of types of extensions of the SM. It is therefore important to reconsider the "problem" in more detail.
One of my key points concerns the different meaning a possible hierarchy problem has in the symmetric and in the broken phase of the SM. In order to understand the point we have to remember why we need the Higgs particle in the SM. The Higgs boson is necessary to get a renormalizable low energy effective electroweak theory [10]. Interestingly, one scalar particle is sufficient to solve the renormalizability problems arising from each of many different massive fields in the SM, of which each causes the problem independently of the others. The point is that this one particle has to exhibit as many new forces as there are individual massive states [11]. All required new interactions are in accordance with the SM symmetry structure in the symmetric phase as we know. The taming of the high energy behavior of course requires the Higgs boson to have a mass in the ballpark of the other given heavier SM states, if it would be much heavier it would not serve its purpose in the low energy regime. It would lead to the so called "delayed unitarity" phenomenon [115]. Note that the Higgs boson has to cure the unphysical mass effects for the given gauge boson masses M W , M Z and fermion masses M f via adequate Higgs exchange forces, where the coupling strength is proportional to the mass of the massive field coupled. A very heavy Higgs boson eventually would decouple and thus miss to restore renormalizability of the massive vector-boson gauge theory. Interestingly, in the symmetric phase the SM gauge-boson plus chiral fermions sector is renormalizable without the Higgs-boson and Yukawa sectors and scalars are not required at all to cure the high energy behavior, because it is renormalizable by its own structure. Therefore, in the symmetric phase the mass degenerate Higgs fields in the complex Higgs doublet can be as heavy as we like. Since unprotected by any symmetry, naturally we would expect the Higgs particles indeed to be very heavy. In fact the "origin" of the Higgs mass is very different in the broken phase, where all the masses, including the Higgs mass itself, are generated by the Higgs mechanism [48,49]. This we learn from the relation m 2 H = 1 3 λ v 2 , holding in the broken phase. In the symmetric phase the effective Higgs mass is dynamically generated by the Planck medium, as we will argue below. Therefore, the usual claim that the SM requires to be extended in such a way that quadratic divergences are absent has no foundation. Purely formal arguments based on perturbative counterterm adjustments do not lead any further.
The hierarchy problem in particular addresses the presence of quadratic UV divergences related to the SM Higgs mass term. Infinities in physical theories are the result of idealizations and show up as singularities in a formalism or in models. UV singularities in general plague the precise definition as well as concrete calculations in quantum field theories (QFT) . 12 A closer look usually reveals infinities to parametrize our ignorance or mark the limitations of our understanding or knowledge. One particular consequence of UV divergences in local QFTs is that the vacuum energy is ill-defined as it is associated with quartically divergent quantum fluctuations. This is another indication that tells us that local continuum QFT has its limitation and that the need for regularization is actually the need to look at the true system behind it. In fact the cutoff system is more physical and does not share the problems with infinities resulting from the idealization realized in the large cutoff limit or lattice continuum limit. In any case the framework of a renormalizable QFT, which has been extremely successful in particle physics up to highest accessible energies, is not able to give answers to the questions related to vacuum energy and hence to all questions related to dark energy, accelerated expansion and inflation of the universe.
Since the SM exhibits non-AF couplings like the U (1) Y coupling g 1 or the Higgs self-coupling λ at scales beyond the zero of the β λ function, also lattice calculations [120][121][122][123][124] strongly suggest that in fact the theory requires a finite cutoff, because the continuum limit at infinite cutoff would be trivial.
It is thus natural to consider the SM to be what we observe as the Low Energy Effective SM (LEESM), the renormalizable tail of the real cutoff system sitting at the Planck scale. As a consequence all properties required by renormalizability, gauge symmetries, chiral symmetry, anomaly cancellation and the related fermion family structure, as well as the existence of an elementary scalar, the Higgs boson, naturally emerge as a consequence of the low energy expansion. 13 We remind that the emergence of SM structures in a low energy expansion is a well investigated subject (see e.g. [125] and references therein). It is often advocated as a tree-unitary requirement but is easily reinterpretable as a low energy expansion where non-renormalizable effects are suppressed by inverse powers in the cutoff. These mechanisms are calculable 12 Taming the infinities we encounter in the theory of elementary particles, i.e. of quantum field theories, by completing them with a cutoff, often called the UV-completion of a QFT, is as old as QFT itself. Actually, it took 20 years from Dirac 1928 (Dirac hole theory of relativistic electron-photon interaction [pre-QED]) to Feynman, Schwinger and Tomonaga in 1948 who found how to deal with the large cutoff limit and making QED a predictive theory. For non-Abelian gauge theories proposed by Yang and Mills in 1954 [116] it took another 17 years until a renormalizable formulation was found by 't Hooft in 1971 [117][118][119] (actually by circumventing a cutoff regularization). 13 It is interesting to note that statistical mechanical systems with short-range exchange and long-range multipole interactions exhibit vector bosons and graviton modes that follow from a multipole expansion of a static potential [126]. In this sense the emergence of gauge-bosons looks pretty natural. within perturbation theory [126][127][128][129][130][131][132]. As SM perturbation theory works at the Z mass scale and gets better with increasing energy these perturbative derivations of gauge symmetry and Higgs structure attain the status of proofs. The infinite tower of higher order operators is suppressed to be invisible. Only a few operators are non-irrelevant and effectively observable, and this is what makes the world look much simpler, than a possibly chaotic Planck medium. In reality infinities related to the relevant operators are replaced by eventually very large but finite numbers, and I will show that sometimes such huge effects are needed in order to understand the real world. I will argue that cutoff enhanced effects are responsible for triggering the Higgs mechanism not very far below the Planck scale and the inflation of the early universe, as outlined already in Sect. 1.
The history of our universe we can trace back 13.8 billion years close to the Big Bang, when the expansion of the universe was ignited in a "fireball", an extremely hot and dense state when all structures and in the end all atoms, nuclei and nucleons were disintegrated into a world of elementary particles only. So the SM provides the key information for what has happened in the early universe, and high energy accelerator experiments are testing processes that only took place in nature in the early history of the universe. If the Higgs boson is the source of dark energy that triggered inflation, its discovery could mark a milestone in our understanding of the dynamics of the very early universe. The origin of cold dark matter remains a mystery, which can have many different explanations.
I think that questions concerning the early universe can be addressed only within a LEESM "extension" of the SM, e.g. given by the SM supplied with a cutoff structure in a minimal way. As we know, in a renormalizable QFT all renormalized quantities as a function of the renormalized parameters and fields in the limit of a large cutoff are finite and devoid of any cutoff relicts! Here we should remember the Bogoliubov-Parasiuk renormalization theorem that states that order by order in perturbation theory the renormalized Green's functions and matrix elements of the scattering matrix (Smatrix) are free of ultraviolet divergences. The theorem specifies a concrete procedure (the Bogoliubov-Parasiuk R-operation) for the subtraction of divergences, establishes the correctness of this procedure, and guarantees the uniqueness of the obtained results, modulo reparametrizations, which are controlled by the renormalization group. In other words, in the low energy world cutoff effects are not accessible to experiments. Consequently, the hierarchy problem cannot be addressed within the renormalizable, renormalized SM, which encodes all observables. In this framework all independent parameters are free and have to be supplied by experiments. In this sense, within the renormalized QFT the hierarchy problem is a pseudo-problem.
To my knowledge the only non-perturbative definition of a renormalizable local quantum field theory is the possibility to put in on a lattice by discretization of spacetime. This again may be taken as an indication that the need for a cutoff actually is an indication that the cutoff exists in the real world. In this sense lattice QFT is closer to the true system than its continuum tail. Of course, there are many ways to introduce a cutoff and actually we cannot know what the cutoff system looks like truly. This is not a real problem if we are interested in the long range patterns mainly. The only thing we have to take care of is that the underlying system is in the universality class of the SM. This in particular concerns the observable degrees of freedom and the emergent symmetries at work, which require the particles to be grouped predominantly in the simplest (lowest dimensional) representation of the corresponding symmetry groups. The simplest symmetry groups with singlets, doublets and triplets are the most natural ones to emerge, as realized within the SM's U (1) Y ⊗ SU(2) L ⊗ SU(3) c gauge symmetry pattern [127][128][129][130][131][132]. More on how the SM may emerge the reader may find in the Appendix.

The Hierarchy Problem Revisited
In [133] already, I outlined the flaws I see in the common reasoning concerning the hierarchy issue. As argued above, I am addressing the hierarchy problem within the LEESM "extension" of the SM. Specifically, I have in mind an implementation of the SM on a Planck lattice (see e.g. [134,135]). The only important point is that we can perform a low energy expansion in the corresponding cutoff. It is an accepted fact that the SM predicts a huge gap between the renormalized and the bare Higgs boson mass. From the LEESM point of view this prediction is what promotes the Higgs boson to be a promising candidate for the inflaton. The hierarchy gap showing up is not something we have to avoid. Now, would-be infinities are replaced by eventually very large but finite numbers, and I will show that sometimes such huge effects provide what we need to understand established phenomena like inflation.
One thing we should remind here: the bare suitably regularized theory has always been the true one. Renormalization always has just been a reparametrization. The bare theory, assumed to exhibit a cutoff of some sort, shows a cutoff dependent large-cutoff tail, sometimes called "preasymptote" [1,114], which is equivalent to a renormalized local QFT in the universality class of the cutoff-system. Thereby it is not important that the bare cutoff system exhibits all symmetries the long-range tail will have, because most of the symmetries of the LEET are emergent. In fact, by a reparametrization of parameters and fields of the preasymptotic theory (renormalizable tail) the residual cutoff-dependence is completely removable (see [114] and references therein). Because the renormalized tail has lost all information about the cutoff, it is nonsensical to say that in the LEET we would naturally expect the Higgs mass to be of the order of the cutoff.
However, in the LEESM "extension" of the SM, bare parameters turn into physical parameters of the underlying cutoff-system being the "true world" at short distances. Then the hierarchy problem is the problem of "tuning to criticality", which concerns the dim < 4 relevant operators, in particular the mass terms. In the symmetric phase of the SM there is only one mass to be renormalized, the others being forbidden by the known chiral and gauge symmetries. For the Higgs field mass which appears in the Higgs potential the fine-tuning to criticality has the familiar form with a coefficient typically C = O (1). To keep the renormalized mass at some small value, which can be seen at low energy, formally m 2 0 has to be adjusted to compensate the huge number δm 2 revealed by the perturbative SM calculation such that about 35 digits must be adjusted in order to get the observed value below the electroweak scale. Is this a problem?
One thing is obvious: our fine-tuning relation exhibits quantities (in the LEESM all observable in principle) at very different scales, the renormalized ones at low energy and the bare ones when approaching the Planck scale. As long as we have no direct access to the Planck physics there is no proven conflict.
Actually, a closer look reveals that in the Higgs phase there is no hierarchy problem in the SM! Why? It is true that in the relation (16) both m 2 0 and δm 2 formally may be expected many many orders of magnitude larger than m 2 . Even worse, in the broken phase δm 2 has a huge negative value and hence m 2 0 must be tuned to be huge negative as well. However, in the broken phase Since v is the result of spontaneous symmetry breaking (non-symmetric ground state) it is per se a low energy parameter related to the emergence of long-range order. Thus in the broken phase the Higgs boson naturally is expected to be light. That the Higgs mass likely is O(M Pl ) in the symmetric phase is what realistic inflation scenarios are demanding.
In the broken phase, characterized by the non-vanishing Higgs field VEV v(μ) = 0, all the masses are determined by the well-known mass coupling relations Here we consider the parameters in the MS renormalization scheme, μ is the MS renormalization scale, which we have to identify with the energy scale of the physical processes or equivalently with the corresponding temperature in the evolution of the universe. The RG equation for v 2 (μ 2 ) follows from the RG equations for masses and massless coupling constants using one of these relations. The evolution of the MS versions of m and v are shown in Fig. 5. As a key relation we use [80][81][82] where γ m 2 ≡ μ 2 d dμ 2 ln m 2 and β λ ≡ μ 2 d dμ 2 λ . We write the Higgs potential as V = m 2 2 H 2 + λ 24 H 4 , which fixes our normalization of the Higgs self-coupling. When the m 2 -term changes sign and λ stays positive, we know we have a first order phase transition (see below). Funny enough, the Higgs particle gets its mass from its interaction with its own condensate! and thus gets a mass in the same way and in the same ballpark as the heavier SM species, which couple strongest to the Higgs field. As mentioned before the Higgs mass cannot be much heavier than the other heavier particles if renormalizability is to be effective at low and moderate energies. H is huge negative, which requires a non-perturbative vacuum rearrangement revealing the mentioned mass coupling relations in terms of a renormalized effective v also for the Higgs particle. Since v ≡ 0 above the EW phase transition point, it makes no sense to say that one naturally has to expect v(μ = M Pl ) = O(M Pl ) . The Higgs-field VEV v is an order parameter, related to the spontaneous breaking of the discrete symmetry 14 Z 2 : H ↔ −H , and is resulting from long range collective behavior. It can be as small as we like. Its value is a function of the effective temperature (energy scale) with its maximum at T = 0, monotonically decreasing with increasing temperature and vanishing at the second order phase transition point T c above which v(T ) ≡ 0 vanishes identically (non-analyticity).
A well known prototype for long range order is the magnetization in a ferromagnetic spin system 15 illustrated in Fig. 6. 14 In the unitary gauge we can avoid problems related to Elitzur's theorem [136], which claims that an order parameter cannot be associated with SSB of a non-Abelian gauge theory. In a physical gauge, on physical Hilbert space, Higgs ghost fields are absent and a Mexican hat potential is a phantom as it only exist if ghost space is taken into the display. A physical Mexican hat potential would imply the existence of three Nambu-Goldstone bosons. 15 As an example we may consider an Ising ferromagnet in D = 2 dimensions, J is the nearest neighbor (n.n.) spin coupling between the spins on a lattice The analogy shows us that v/M Pl 1 is not unnatural since v = 0 emerges only below a critical temperature, which is not in a simple way related to M Pl . The EW scale is set by v(μ) and depends on μ. As we learn from Fig. 5, at low energy v(0) = 1/( √ 2G μ ) 1/2 ≈ 246 GeV, but interestingly, in contrast to the magnetization of a ferromagnet, v(μ) is increasing rather than monotonically decreasing with increasing μ. This is because of the rich conspiring dynamics of the SM encoded in the RG equation (18) 16 , yet v(μ) is vanishing at μ 0 ∼ 10 16 GeV: v(μ) → 0 when μ < → μ 0 , as we will see later. The second order phase transition (PT) point is a point of nonanalyticity i.e. exhibits singular behavior and physics in the ordered phase and the disordered phase are very different.
Considering a ferromagnet one has to tune the temperature T to criticality in order to find the PT point. What is tuning the temperature to criticality in the SM? The answer is the expansion of the universe, which provides a scan in temperature (see also [137]). The maximum value of v(μ) is achieved in the low energy limit at μ = 0. Why should the magnitude of v(0) be of the order of the Planck scale, given the fact that above the phase transition point, in the disordered phase, the VEV is actually vanishing identically?
This shows that the Higgs boson mass renormalization equation is not a static equation but is subject to a sophisticated dynamics driven by the expansion of the universe.
In the symmetric phase at very high energy we see the bare system. There the Higgs field is a collective field exhibiting an effective mass generated by radiative effects within the Planck system such that m 2 0 ≈ δm 2 at M Pl . In particle physics a depending on temperature T and n.n. spin interaction strength J . For more details see e.g. [113] 16 Fig. 5 shows that the Higgs mass parameter m is varying little in the broken phase, while v = √ 6/λ m increases substantially because λ decreases rapidly. radiatively induced mass is known from the Coleman-Weinberg mechanism [138], now in the symmetric phase and applied to the Planck medium. Such mechanism, which is natural in this context, eliminates a possible fine-tuning problem at all scales. There are many examples in condensed matter systems, like the effective mass of the photon in the superconducting phase (Meissner effect) or the effective mass of the effective field which encodes the spin-singlet electron pairs (Cooper pairs) in the Ginzburg-Landau (GL) model [139] of superconductivity. 17 The latter directly corresponds to the Abelian Higgs model. Emerging as an effective field from the hot Planck system, which exhibits all types of excitations, it is also pretty obvious that the Higgs field couples to all these modes that we see as Yukawa and Higgs to gauge boson couplings in the SM. That these couplings exhibit the symmetries of the SM is again due to the fact that only the renormalizable tail can be seen at low energies. All Planck system modes that do not conspire, as SM degrees of freedom and their couplings do within the SM, are not perceivable at long distances. The SM emerges as a self-organized system.
On the one hand we know that astronomy and astrophysics are unthinkable without the input from laboratory physics in general and particle physics in particular. On the other hand, it is not new that particle physics is learning from cosmology. What is required to explain inflation, baryogenesis, nucleosynthesis, CMB patterns, dark matter, etc. ? If the SM has an extrapolation up to the Planck scale, evidently one is able to confront SM predictions with physics established to have happened in the early universe. In contrast to the old paradigm of an empty vacuum: we know that the ground state of the world is filled with dark energy, with a Higgs condensate and quark and gluon condensates. All these effects have been showing up at certain times and play a key role in the evolution of the universe. Obviously, there are plenty of questions to be answered in order to get a better understanding of how the universe has been shaped after the Big Bang.

Running SM Parameters Trigger the Higgs Mechanism
In Sect. 2.2 already, we have discussed how the Higgs boson discovery has been revealing a peculiar value for the Higgs boson self-coupling, which largely clarified the path of extrapolating the SM to higher energies. We remind that all dimensionless couplings satisfy the same RG equations in the broken and in the unbroken phase and are not affected by any power cutoff dependencies. This is as it has to be because the Higgs mechanism (SSB) does not alter the UV behavior. The evolution of the SM couplings in the MS scheme up to the Planck scale has been investigated in [56,[65][66][67][68][69][70][71][72][73][74][75][76][77]85,144,145], and has been extended to include the Higgs-field VEV and the masses in [1,83,84]. 17 Originally the Ginzburg-Landau theory of superconductivity has been proposed as a macroscopic phenomenological effective theory describing type-I superconductors without reference to microscopic properties. Later Bardeen-Cooper-Schrieffer could explain superconductivity from its microscopic structure in their BCS-theory [140,141]. Afterward Gor'kov derived the GL-theory [142] showing that in some limit all GL parameters have a microscopic interpretation. In addition, Abrikosov showed that GL-theory also models type-II superconductors [143]. The effective GL-theory thus efficiently describes a rich variety if superconducting systems, without the need for a detailed microscopic understanding.
Except for g 1 , which increases very moderately, all other couplings decrease and stay positive up to the Planck scale. This strengthens the reliability of perturbative arguments and reveals a stable Higgs potential up to the Planck scale [1,83,84]. While most analyses [65,68,69,85,144,145] find that for the given Higgs mass value range vacuum stability is nearby only (meta-stability), 18 and the SM actually fails to persist up to the Planck scale, our evaluation of the matching conditions yields initial MS parameters at the Z boson mass scale which evolve preserving the positivity of λ. Thereby the critical parameter is the top-quark Yukawa coupling, for which we find a slightly lower value, which is based on the analysis [83,84]. My MS input at M Z is [1] g 3 = 1.2200, g 2 = 0.6530, g 1 = 0.3497, y t = 0.9347 and λ = 0.8070. At M Pl I get g 3 = 0.4886, g 2 = 0.5068, g 1 = 0.4589, y t = 0.3510 and λ = 0.1405 (see Table 1). In view of the fact that the precise meaning of the experimentally extracted value of the top-quark mass is not free of ambiguities, usually it is identified with the on-shell mass M t (see e.g. [69,83,84,146] and references therein), it may be premature to claim that instability of the SM Higgs potential is a proven fact already [86]. As I have elaborated in Sect. 2.1, the implementation of the matching conditions is not free of ambiguities, while the evolution of the couplings over many orders of magnitude is rather sensitive to the precise values of the initial couplings. Accordingly, all numbers presented in this article depend on the specific input parameters adopted, as specified in [1,83,84]. In case the Higgs self-coupling has a zero λ(μ 2 ) = 0, at some critical scale μ c below M Pl , we learn from Eq. (18), or more directly from v(μ 2 ) = 6m 2 (μ 2 )/λ(μ 2 ) λ→+0 → ∞that the SM loses its "being well-defined" above this singular non-analytic point. 19 For our input parameters Table 1 shows that the relevant running MS parameters at Planck scale are of comparable size in the range 0.51 for g 2 being the largest here and 0.35 for y t being the smallest, with √ λ at 0.375 slightly larger in our normalization. It tells us that approximations like the gaugeless limit (g 1 = g 2 = 0) or assuming λ ≈ 0 relative to other couplings are not viable approximations near M Pl .
For what follows we take up what Shaposhnikov et al. [39] say about vacuum stability in their conclusion: "Although the present experimental data are perfectly consistent with the absolute stability of the Standard Model within the experimental and theoretical uncertainties, one should not exclude the possibility that other 18 Most groups are adopting essentially the same input parameters presented in [65,68,85] and a radiatively corrected effective potential and find the vacuum to lose stability at about a surprisingly low scale of about μ ∼ 10 9 GeV [input not independent]. Keep in mind: the Higgs boson mass miraculously turns out to have a value very close to what was expected from vacuum stability. It looks like a tricky conspiracy with other couplings to reach this "purpose". Assuming vacuum stability, the narrow stability window actually makes the Higgs mass to be a predictable quantity if we consider the other SM parameters as given. Also imposing Planck scale boundary conditions may be argued to fix the Higgs boson mass [66,67]. If the Higgs boson misses to stabilize the vacuum, why does it just miss it almost not? 19 As we have argued earlier we consider the bare Higgs potential to be the true potential, except that the bare parameters have to be calculated bottom-up from the known values at low energy. A low energy reparametrization also affects the form of the potential by radiative corrections as we know from Coleman-Weinberg [138]. The correspondingly modified effective potential plays a crucial role when the potential gets unstable and actually can turn instability into meta-stability [39,65]. This will be discussed in Sect. 5.1 below. The Planck medium, from which the SM derives as a long distance tail, certainly exhibits a stable ground state. This we infer from our mere existence. Note that λ = 0 is an essential singularity and the theory cannot be extended beyond a possible zero of λ. Remind that v = 6m 2 /λ i.e. v(λ) → ∞ as λ → 0. Besides the Higgs boson mass m H = √ 2 m all masses m i ∝ g i v → ∞ would yield a different cosmology experiments will be able to establish the meta-stability of the electroweak vacuum in the future." But, based on a slightly modified evaluation of MS parameters [83,84] (which revealed vacuum stability), we adopt the view: "Although other evaluations of the matching conditions seem to favor the meta-stability of the electroweak vacuum within the experimental and theoretical uncertainties, one should not exclude the possibility that other experiments and improved matching conditions will be able to establish the absolute stability of the Standard Model in the future." Running couplings can affect dramatically the quadratic divergences and the interpretation of the hierarchy problem. Quadratic divergences have been investigated at one-loop in [4] (see also [5,147,148]), at two loops in [6][7][8]. At n loops the quadratic cutoff-dependence is of the form where the n-loop coefficient only depends on the gauge couplings g 1 , g 2 , g 3 , the Yukawa couplings y f and the Higgs self-coupling λ. Neglecting the numerically insignificant light fermion contributions, the one-loop coefficient function C 1 may be written as and is uniquely determined by dimensionless couplings. The latter are not affected by quadratic divergences such that standard RG equations apply. Surprisingly, as first pointed out by Hamada, Kawai and Oda in [7], taking into account the running of the SM couplings, the coefficient of the quadratic divergences of the bare Higgs mass correction can vanish at some scale, given the specific SM couplings that became available after the Higgs boson discovery. In contrast to our evaluation Hamada et al. actually find the zero to lie above the Planck scale, having adopted input MS parameters from [65]. In our analysis, relying on matching conditions for the topquark mass analyzed in [83,84], we get a scenario where λ(μ 2 ) stays positive up to the Planck scale and looking at the relation between the bare and the renormalized Higgs mass we find C 1 and hence the Higgs mass counterterm to vanish at about μ 0 ∼ 1.4 × 10 16 GeV, not very far below the Planck scale. The next-order correction, first calculated in [6,8] and confirmed in [7] reads and numerically does not change the one-loop result significantly. The same results apply for the Higgs potential parameter m 2 , which corresponds to m 2= 1 2 m 2 H in the broken phase. For scales μ < μ 0 we have δm 2 large negative, which is triggering spontaneous symmetry breaking by a negative bare mass m 2 0 = m 2 + δm 2 , where m again denotes the renormalized mass. The phase transition is illustrated in Fig. 7.
The jump taking place here in the vacuum energy is given by 20 As a CC contribution it is of negative sign and 50 orders of magnitude off relative to what corresponds to the observed Λ CMB (see also [23]). However, the effect is small , which will be discussed in Sect. 6. At μ = μ 0 we have δm 2 = 0 and the sign of δm 2 flips, implying a phase transition to the symmetric phase. Finite temperature effects [149][150][151][152], which must be included in a realistic scenario, turn out not to change the gross features of our scenario, unless μ 0 would turn out to lie much closer to Λ Pl [1]. A different effect is due to the change in the effective mass resulting from the Wick reordering of the Lagrangian by a nonvanishing Φ + Φ . This will be discussed in Sect. 6. It produces a larger shift of the What do we learn from this analysis? The Higgs mechanism is dynamically triggered as temperature in the universe drops below μ 0 . In the low energy phase the Higgs boson mass M H substitutes √ 2m and in fact has to be calculated from the vacuum rearrangement (see Fig. 12). Now m H turns into an emergent mass, which is determined by the mass-coupling relation (17) like for all other massive particles in the Higgs phase. At the transition point we have δm 2 = 0 and no hierarchy problem. While above μ 0 the shift δm 2 is physical and emergent from the interaction in the Planck medium, below μ 0 the shift δm 2 looses its physical meaning. This is because at μ = μ 0 the enhanced cutoff-effects are nullified. At this point the access to cutoff effects gets lost and we enter the renormalizable renormalized low energy phase. Below μ 0 we still use δm 2 H in perturbative mass renormalization, where it is now large negative if we still insist on using the now physics-wise inaccessible Planck scale as a UV cutoff. I would say that argumentations based on (16) now turn into formal nonsense. Not only the magnitude of the cancellation is arbitrary, but it also has the wrong sign, for what could be related to a physical mass. The physical mass is determined by the curvature at the minimum of the potential. Key outcome of our calculation is the observation that the SM at high enough sub-Planckian energies undergoes a transition into the symmetric phase [1], presuming a stable vacuum. Fig. 1 displays the SM prediction for the effective Higgs mass as a function of the energy scale. This profile promotes the Higgs boson to act as an inflaton as discussed in Sect. 1 already.

Vacuum Stability and Effective Potential
The classical Higgs potential (2) for λ > 0 is bounded from below and has a trivial minimum for m 2 > 0 at φ 0 = 0, and a non-trivial minimum at φ 2 0 = −6m 2 λ for m 2 < 0. When the classic potential turns unstable, because λ is running to be negative, the analysis of the vacuum stability has to be based on the effective potential, which is obtained by including the quantum corrections [138,154,155]. 21 The effective potential is gauge-and scale-dependent and not an observable. In the Landau gauge and the MS scheme it can be written as [156][157][158][159] (also see [57,58] with where κ = 1/(4π) 2 and m i are the masses of different particles in the background of the classical Higgs source field φ c of the generating functional for the irreducible 21 Being a part of the SM Lagrangian the Higgs potential term considered so far gets reparametrized by a change of the effective parameters and the effective Higgs field and by appropriate counterterms only, as long as perturbation theory does not break down. All perturbative physics is obtained as usual by means of the renormalizable Lagrangian, written in terms of the quantized fields, and the corresponding Feynman rules. Also note that the Higgs contribution to the energy-momentum tensor of Einstein gravity is represented by the symmetric energy-momentum tensor in terms of the Higgs part of the bare SM Lagrangian. 22 As shown in [138], the potential satisfies the RG equation where λ i = m 2 , λ, g = g 1 , g = g 2 , g s = g 3 , y t with corresponding beta-functions β i and γ the anomalous dimension of the Higgs field. The RG as usual is solved along characteristic curves where t parametrizes the position on the curve. The solution reads

Higgs vertex functions, which upon renormalization is given by
The effective potential as derived in the symmetric phase include the would-be Higgs ghosts G contribution as physical degrees of freedom, in the broken phase Higgs ghosts are massless in the Landau gauge (would-be Nambu-Goldstone bosons). In the symmetric phase they contribute as three additional Higgs particles. As we know the Higgs boson mass in the broken phase ( where v refers to the EW vacuum. Two-loop corrections V 2 have been calculated in [157,158] and may be found in more condensed form in [65]. V 3 includes the leading three-loop corrections computed in [160]. The remainder V rem represents the higher-order contributions, which include also the higher dimension operators starting at four loops [158,161,162]: where L is number of loops. The wavefunction renormalization of the Higgs field takes the form where γ (t) = d ln Z φ /dt is the anomalous dimension of the Higgs field: Finally, the scale μ(t) is related to the running parameter t by where μ is a fixed scale, that we will take equal to the physical top-quark mass, M t as a reference point. Observable physical predictions up to perturbative truncation errors do not depend on the choice of the renormalization scale. This can be used in order to keep radiative corrections moderate by choosing μ(t) = φ(t) which avoids large logarithms at any given t, since then ln m 2 W (t)/μ 2 (t) = ln g 2 (t)/4 etc. (see [158]). One also may choose μ(t) = φ c in which case ln m 2 W (t)/μ 2 (t) = ln g 2 (t)/4+2Γ etc. The correction Γ = φ M t γ (μ) d ln(μ) stems from the field renormalization factor Z φ .
As elaborated in [57,58] for high Higgs fields the effective potential may be cast into the simple form where it is dominated by the quartic term and λ(φ) depends on φ the same as the running coupling λ(μ) depends on the running scale μ = φ c with modified coupling [85] λ eff ≈ λ + κ 9 4 g 4 2 ln up to less relevant corrections. The crucial point is that for parameters as [Deg] in Table 1 the correction term is positive all up to the Planck scale. At the EW scale the leading positive λ-term dominates λ eff up to scales where λ approaches a zero and there changes sign, if such a zero exists, which depend on the precise input values at the EW scale. In the vicinity above the zero of λ actually λ eff remains positive and such stabilizes the Higgs vacuum to somewhat higher scales but also turns negative to a metastable state before reaching the Planck scale (see Fig. 3 in [65]). In contrast for the parameter set [Jeg] the correction Δλ eff also starts positive but at higher scales takes negative values. These are small, however, and are not affecting the positivity of λ eff itself as seen in the left panel of Fig. 9. In the stable vacuum scenario radiative corrections of the effective potential are moderate and do not affect the main pattern as long as λ remains positive. The cutoff power enhanced effects are always much larger than the standard radiative corrections to the effective potential, provided the former are taken into account. This we will have to remind also for the discussion to follow in Sect. 6. The quantum corrections modify the shape of the potential such that a second minimum at some higher (Planck) scale may be induced (see Fig. 10 and Fig. 7 in [65]). As first discussed in [158], a second minimum is also obtained when a transient instability emerges above our Higgs transition point μ 0 when the bare mass term gets positive and actually gets huge because of the quadratic cutoff-enhancement. For especially fine-tuned parameters this may happen also if radiative corrections are not yet included. In any case the existence of a second minimum depends significantly on the higher order corrections. Depending on the values of the Higgs boson and topquark masses the lifetime of the EW vacuum can be larger or smaller than the age of the Universe. The first case corresponds to the metastable scenario.  Table 1. Left: the effective Higgs self-coupling λ eff governing the SM effective potential V eff ∼ λ eff 24 φ 4 for large fields. "CEQ" is one-loop improved from [57,58], "Deg" is two-loop improved from [65]. The correction Δλ eff represents the corrections included in "Deg" relative to λ bare . Right: the bare potential compared with different approximations of the effective potential: one-loop improved "Deg-1", two-loop improved "Deg-2" and three-loop improved "Mar-3" [160] Fig. 10 The form of the effective potential for the Higgs field φ which corresponds to the stable, critical and metastable electroweak vacuum. The pattern displayed admitting for two minima at non-zero field values requires the effective potential to exhibit even powers of φ up to φ 8 . v is the location of the EW minimum and φ min v is the value of a new minimum For stationary points φ 0 much larger than the electroweak scale one has λ eff 1 and the curvature of the potential is given by [59] Therefore, in order that the potential exhibits a second minimum, the function β λ must have passed a zero, because we know that β λ is negative at EW scales. For the parameter set [Jeg] a zero is found at about μ λ 10 17 GeV. What happens for the two parameter sets [Jeg] and [Deg] is shown in Fig. 11, which also illustrates the significance of the radiative corrections. In the stability case the effective potential does not alter the main picture, while in the metastable case a second minimum is also missing and the potential turns unbounded from below way below the Planck regime. Since the tunneling rate to the Planck regime is exceedingly low, the EW vacuum still looks to be stable. As follows from the SM RG, because β λ contains parts which are not proportional to λ, the Higgs self-coupling λ is the only SM dimensionless coupling that can change the sign with increasing energy scale.
In our case, where λ(μ) > 0 up to M Pl , in the early phase of the expanding universe, the effective potential is approximated by and the correction turns out not to be significant for what concerns the scenario as such. Actually, the upshot of the two-loop analysis in [158] has been that "the requirement that the electroweak vacuum remains stable turns out to be essentially identical to the requirement that λ remains positive". One has to keep in mind that a metastable EW ground-state in a globally unstable potential, as found in commonly accepted analyses [39,65,85], very likely does not model what truly happens at the Planck scale. It could signal the need for an extension of the SM including new physics or that the analysis underestimates uncertainties.

The Cosmological Constant: Dark Energy Provided by the Higgs Scalar
It is crucial that in the early universe both terms in the Higgs potential are positive in order to condition slow-roll inflation during long enough time. In fact the quadratically and quartically cutoff enhanced terms in the Higgs potential enforce the condition 2 + V (φ), we arrive at the equation of state w = p/ρ ≈ −1 characteristic for dark energy and the equivalent CC (see e.g. [21,[163][164][165] and references therein). A first remarkably precise measurement of the dark energy equation of state w = −1.01 ± 0.04, has been obtained by the Planck mission [17,18] recently (for an actual review see [166]). A more detailed study [2] shows that the enhanced Higgs boson effective mass alone actually does not provide a sufficient amount of inflation, which is required to inflate the causal CMB cone to include the full CMB sky. 23 One important quantity we have not taken into account so far is the vacuum energy A key point is that in the LEESM scenario the vacuum energy is a calculable quantity. In the symmetric phase SU(2) symmetry implies that while Φ(x) ≡ 0 the composite field Φ + Φ(x) is a singlet such that the invariant vacuum energy is given just by simple Higgs field loops H 2 =: ; This provides a CC given by A Wick ordering type of rearrangement of the Lagrangian also leads to a shift oft the effective mass For our values of the MS input parameters the zero in the Higgs mass counter term and hence the phase transition point gets shifted downwards as follows The shift is shown in the right panel of Fig. 8. We notice that the SM predicts a huge CC at M Pl : exhibiting a very weak scale dependence (running couplings) and we are confronted with the question how to get rid of this huge quasi-constant? Remember that ρ φ has no direct dependence on a(t). An intriguing structure again solves the puzzle. The effective CC counterterm has a zero, which again is a point where renormalized and bare quantities are in agreement: with X (μ) 1 8 (2C(μ) + λ(μ)) which has a zero close to the zero of C(μ) when 2 C(μ) = −λ(μ). Note that C(μ) = −λ(μ) is the shifted Higgs transition point.
Again we find a matching point ρ Λ0 = ρ Λ between the low-energy and the highenergy world. At this point the memory of the quartic Planck scale enhancement gets lost, as it should be since we know that the low energy phase does not provide access to cutoff-effects.
Crucial point is that acquires positive bosonic contribution and negative fermionic ones, with different scale dependence. 24 X can change a lot (pass a zero), while individual couplings are weakly scale dependent with y t (M Z )/y t (M Pl ) ∼ 2.7 the biggest and g 1 (M Z )/g 1 (M Pl ) ∼ 0.76 the smallest change. Obviously, the energy dependence of any of the individual couplings would by far not be able to sufficiently diminish the originally huge cosmological constant. Only the existence of a zero in the coefficient function X (μ) is able to provide the dramatic reduction of the effective CC by nullifying the huge cutoff-sensitive prefactor. At the Higgs transition point, as soon as m 2 < 0 for μ < μ 0 , the vacuum rearrangement of the Higgs potential takes place. As a result at the minimum φ v of the potential we should get V (0) + V (φ v ) ∼ (0.00171 eV) 4 about the observed value of today's CC (see Fig. 12). How can this be? Indeed, at the zero of X (μ) we have ρ Λ0 = ρ Λ and one may expect that like the Higgs boson mass another free SM parameter is to be fixed by experiment here. 25 One might expect ρ Λ to be naturally small, since the Λ 4 Pl term is nullified at the matching point. Note that the huge cutoff prefactors act as amplifiers of small changes in the effective SM couplings. But how small we should expect the low energy effective CC to be? In fact, in the LEESM scenario neither the Higgs mass nor the CC are really free parameters in the low-energy world. Given the other relevant SM parameters, the Higgs self-coupling has to be constrained to a window where the Higgs potential remains stable up to the Planck scale. Similarly, the originally large CC, which is required to provide a sufficient amount of inflation, has to get tuned down such that inflation ends up at the critical density of a flat universe. The late CC as part of the critical density then only can be a fraction of the latter.

A Self-organized Cosmological Constant?
Implications of inflation we already outlined in Sect. 1 after Eq. (5). Our analysis of the LEESM showed that the CC is very much time dependent especially through the running of the SM parameters and phase transitions taking place in the evolution of the universe (see also the Quintessence scenario advocated in [42][43][44][45]). The typical problem is that in general one gets a CC that is way too big, and this looks to create a tremendous fine-tuning problem. For the SM this concerns the contribution to the vacuum density via the Higgs-field VEV in the broken phase, as well as the contributions from spontaneous breakdown of chiral symmetry, which both are much too big and even of wrong sign. Interestingly, our Higgs inflation scenario predicts a large positive DE, which actually implies that ρ tot ρ crit before inflation sets in. This means that at Planck time k = +1 and Ω k in Eq. (7) evaluated at Planck time is large negative if a(t Pl ) is of Planck size. It is important to keep in mind that in Big Bang cosmology ρ tot at the beginning is always dominated by the radiation density since ρ rad ∝ a(t) −4 grows fastest when a(t) gets smaller as we are going back in time. Because they also scale with inverse powers in a(t), also the matter-density and the curvature terms first overshoot the CC supplied by Higgs boson system. This is possible because ρ Λ ∼ (1.29 M Pl ) 4 is of comparatively moderate size, although extremely big relative to the critical density. However, if inflation is at work, the final vacuum density is fixed, whatever the initial density has been. Given that Ω tot = Ω Λ +Ω DM +Ω BM +Ω rad = 1 with 1 > Ω DM > Ω BM > Ω rad > 0 we know that Ω Λ being positive must be of order Ω tot , actually a fraction of it. As a non-vanishing ρ Λ0 at Planck time is needed, it is not unlikely that the other components contributing to the total energy density do not saturate the bound. Actually we know that normal matter including the tiny radiation density represents about 5% of the critical density only. This means that the fine-tuning is dynamically enforced by inflation and the value of today's dark energy density ρ 0Λ = μ 4 0Λ ; μ 0Λ = 0.00171 eV (41) looks all but exotic. While Ω rad and very likely Ω BM are essentially LEESM predictions if we include the B + L violating dimension 6 four-fermion operators, Ω DM is the only missing piece which remains an open problem and definitely requires additional 0Λ with μ 0Λ 0.00171 eV. Note: ρ Λ (t) includes besides the large positive V (0) also negative contributions from vacuum condensates, like Δρ EW from the Higgs mechanism and Δρ QCD from the chiral phase transition beyond the SM physics. This also concerns contributions from quark-and possible gluon-condensates, which we do not explicitly consider here.
Provided SM parameters indeed support a stable Higgs potential up to M Pl , inflation and the CC itself are SM ingredients leading to a highly self-consistent conspiracy which shapes the universe. Fig. 13 shows the development of the quadratically and the quartically enhanced terms in the symmetric phase of the SM, and its matching to the low energy world.

Inflation and Reheating
In contrast to standard scenarios of modeling the evolution of the early universe, SM cosmology is characterized by the fact that almost everything is known, within uncertainties of the parameters and perturbative approximations. In LEESM cosmology the form of the potential is given by the bare Higgs potential V (φ) = m 2 2 φ 2 + λ 24 φ 4 as part of the SM Lagrangian, the parameters are known, calculable in terms of the low energy parameters, the only unknown is the magnitude of the Higgs field. The latter must be large-trans-Planckian-in order to get the required number of e-folds N given by The second form is obtained using the field Eq. (4). Note that N is determined entirely by the scalar potential. Needed is N 60 in order to cover the CMB causal cone. By definition, exp N is the expansion factor a(t e )/a(t i ) = exp H (t e − t i ), where a(t) is the Friedmann-Robertson-Walker radius of the universe at cosmic time t, t i denotes the begin of inflation and t e the end of inflation and H the Hubble constant. For our set of MS input parameters we require φ 0 = φ(μ = M Pl ) ≈ 4.5 M Pl . Shortly 2 is well satisfied, by the fact that in the symmetric phase the mass term as well as V (0) = V (φ) are huge and start to dominate quickly. Because of the large initial field strength φ 0 , however, the interaction term is actually dominating for a short time after the initial Planck time t Pl . The field equationφ + 3Hφ = −V (φ) then predicts a dramatic decay of the field, such that in almost no time, still under slow-roll conditions, the mass term dominates and for what follows the field equation predicts an exponential decay followed by harmonic oscillation setting in. The universe thus undergoes an epoch of Gaussian inflation as confirmed by observation [22]. The time evolution is displayed in Fig. 14 and it is very interesting to see which term dominates during which time slice. Obviously, the fast decay of the Higgs field stops inflation (see Fig. 15), in spite of the fact that a CC V (0) persists to be substantial at first.
A highly non-trivial challenge is the calculation of the spectral indices which sensitively depend on the form of the potential and which have been extracted from the observed CMB radiation fluctuation spectrum. From the theory point of view the indices are constrained by the slow-roll criteria ε 1, ensuring p −ρ, and ε, η 1, which ensures that the slow-roll condition hold for long enough time, while maintainingφ 3 Hφ before oscillations start. These conditions also should be satisfied in order to ensure the required amount of inflation. In our case we are confronted with a SM prediction modulo the unknown Higgs field φ 0 = φ(μ = M Pl ). The calculation presented in [2] shows that a prediction of η is delicate, but present uncertainties in predicting the bare Higgs potential at post-inflation times certainly do not allow us the draw definite conclusions. Given that there are many predictions that look to work surprisingly well, I would be surprised if the Higgs boson inflation would not work in predicting also the spectral exponents in agreement with observation at the end.

Reheating by Higgs Boson Decays
In the symmetric phase all four Higgs fields are physical and very heavy and rather unstable. The Yukawa couplings at inflation times are pretty well known and the Higgs bosons decay predominantly (largest Yukawa couplings) into as yet massless top-antitop pairs and lighter fermion-antifermion pairs H , φ 0 → tt, bb, · · · , H + → tb · · · , H − →tb · · · and are thereby reheating the young universe, which just had been cooled down dramatically by inflation. Preheating is suppressed in SM inflation as in the symmetric phase bosonic decay channels like H → W W and H → Z Z are absent at tree level. The CP violating decays H + → td [rate ∝ y t y d V td ] H − → bū [rate ∝ y b y u V ub ] likely are important for baryogenesis. Closely following the Higgs transition, where m 2 in the Higgs potential changes sign, the electroweak phase transition takes place. After it the now heavy top-quarks decay into normal matter as driven by CKM [167,168] couplings and phase space. At these scales the B + L violating dimension 6 operators [169][170][171] can still play a key role for baryogenesis and together with decays like t → de + ν provide CP violating reactions during a phase out of thermal equilibrium. 26 For details see [1,2,133].
A very different model of Higgs inflation, which has barely something in common with our LEESM scenario, is the Minkowski-Zee-Shaposhnikov et al. [33][34][35][36][37][38][39] so called non-minimal SM inflation scenario. It is based on the following points: (i) Einstein gravity has to be extended by adding G μν → G μν + ξ (Φ + Φ) R g μν to Einstein's equation. On the source side the model is assuming the renormalized low energy T μν supplied by the renormalized SM (no relevant operator enhancement). The new term is a direct coupling of the gauge invariant Higgs field singlet operator Φ + Φ to the scalar Ricci curvature R. This extra term violates the equivalence principle, yet so far without observable consequences. (ii) Choose ξ large enough in order to get a sufficient amount of inflation, which requires a rather large value ξ ∼ 10 4 . The entire inflation pattern then essentially depends on ξ only (inflation "by hand"). In case ξ = O(1) the added non-minimal coupling term is tiny and does not affect our LEESM or standard inflation scenarios. (iii) Assume quadratic and quartic SM divergences are absent (argued by their absence in dimensional regularization (DR) and MS renormalization, which is a misleading purely formal argument in my opinion). (iv) Assume the SM to be in the broken phase at Planck scale, which looks unnatural since SSB is a low energy phenomenon, which assumes the symmetry to be restored at the short distance scale! Note: (1) It is well possible maybe even likely that such non-minimal gravity couplings of the Higgs field exist and could play a role when curvature is very high.
However, the coupling ξ would rather be O(1) than fine-tuned to be about ξ ∼ 10 4 .
(2) Dimensional regularization and MS renormalization both are possible in perturbation theory only. There is no corresponding non-perturbative formulation (simulation on a lattice) or measuring prescription (experimental procedure). The MS scheme is based on a finite part prescription (singularities nullified by hand), which can only be used to calculate quantities that do not exhibit any singularities at the end. As elaborated earlier in Sect. 4, the hierarchy problem cannot be addressed within the dimensionally regularized SM or adopting the MS scheme in a renormalized environment. In other words, dimensional renormalization by no means explains the absence of power enhanced terms in a LEET scenario. These terms are there and have to be accounted for.

Remark on Trans-Planckian Higgs Fields
If the SM Higgs boson is the inflaton, sufficient inflation requires trans-Planckian magnitude Higgs fields at the Planck scale. At the cutoff scale the low energy expansion obviously gets obsolete and likely we cannot seriously argue with field monomials and the operator hierarchy appearing in the low energy expansion. What is important is that the field is decaying very fast (see Fig. 15). Formally, given a truncated series of operators in the potential, the highest power is dominating when approaching the trans-Planckian regime. One then expects that for some time the φ 4 term of the potential is dominating, the decay of the field is then exponential, for higher dimensional operators it is faster than exponential, such that the field very rapidly reaches the Planck-and sub-Planck regime. This means that the mass term is dominating after a very short period and before the kinetic term becomes relevant and slow-roll inflation ends. So fears that higher order operators in low energy effective scenarios with trans-Planckian fields would mess up things are unfounded. 27 Obviously, without the precise knowledge of the Planck physics, very close to the Planck scale we never will be able to make a precise prediction of what is happening. This however seems not to be a serious obstacle to quantitatively describe inflation and its properties as far as they can be accessed by observation. The LEESM scenario in principle predicts not only the form of the effective potential not far below the Planck scale but also its parameters and the only quantity not fixed by low energy physics is the magnitude of the field at the Planck scale. We also have shown that taking into account the running of the parameters is mandatory for understanding inflation and reheating and all that.
Trans-Planckian fields are not unnatural in a low energy effective scenario as the Planck medium exhibits a high temperature and temperature fluctuations imply amply of higher excitations, which we may perceive as a chaotic state. While the Planck Fig. 15 The trans-Planckian Higgs field at t Pl decays very fast and inflation gets stopped soon. Left: the decaying Higgs field. Right: the inflating a(t) medium will never be accessible to direct experimental tests, a phenomenological approach to constrain its effective properties is obviously possible, especially by CMB data [172].
In the extremely hot Planckian medium, the Hubble constant in the radiation dominated state with effective number g * (T ) = g B (T ) + Often it is argued that trans-Planckian fields are unnatural in particular in a LEET scenario [173]. I cannot see any argument against strong fields and LEET arguments (ordering operators with respect to a polynomial expansion and their dimension) completely lose their sense when E/Λ Pl 1. As mentioned already, provided the Higgs field decays fast enough, towards the end of inflation we expect the mass term to be dominant such that a Gaussian fluctuation spectrum prevails. The quasi-constant CC V (0) at these times mainly enters the Hubble constant H and does not affect the fluctuation spectrum. As the originally large V (0) get nullified at μ CC also the Hubble constant suffers a jump down to a value as known from standard ΛCDM cosmology.

Remarks for the Skeptic
How do our results depend on the true UV completion? In other words, how realistic are the numbers I have presented?
In order to answer these questions we have to stress once more the extreme size of the cutoff, M Pl >>>> ... from what we can see!, which lets look what we can explore to be ruled by fundamental principles like the Wightman axioms (the "Ten Commandments" of QFT) or extensions of them as they are imposed in deriving the renormalizable SM. In the LEESM approach many things are much more clear-cut than in condensed matter systems, where cutoffs are directly accessible to experiments. In other words, in condensed matter systems the gap between the microscopic structure and its macroscopic manifestation is by far nowhere nearly as big as in our LEESM case. Also lattice QCD simulations differ a lot, as cutoffs are always close-by, such that lattice artifacts affect results throughout before the extrapolation to the continuum has been performed.
We also have to stress that taking actual numbers too serious is premature as long as even the realization of vacuum stability is in question. Detailed results evidently depend sensitively on accurate input values and on the perturbative approximations used for the renormalization group coefficients as well as for the matching relations needed to get the MS input parameters in terms of the physical (on-shell) ones. After all we are attempting to extrapolate over 16 orders of magnitude in the energy scale. Such an attempt may look to be megalomaniacal, but it is a bottom-up approach that appears to lead to a reasonable and a very possible scenario that is able to explain and predict inflation and reheating. And it is a very modest step in relation to attempts to construct a TOE for example.
The next question is how close to M Pl can we trust our extrapolation? It is very important to note that above the EW scale [v ∼ 250 GeV] perturbation theory seems to works the better the closer we are near the Planck-cutoff, vacuum stability presupposed. As long as we are talking about the perturbative regime we can expand perturbative results in powers of E/Λ Pl up to logarithms. Then we have full control over the cutoffdependence to order O((E/Λ Pl ) 2 ), corresponding to dim ≥ 6 operator corrections. Effects O((E/Λ Pl )), related to dim 5 operators, only show up in special circumstances e.g. in scenarios related to generating neutrino masses and mixings and the see-saw mechanism.
The true problem comes about when we approach the Planck scale, where the expansion in E/Λ Pl completely breaks down. Especially, it does not make sense to talk about a tower of operators of increasing dimensions. This does not mean that everything gets out of control. If the "ether" would be something that can be modeled by a lattice SM, implemented similar to lattice QCD, one could still make useful predictions, which eventually could be tested in cosmological phenomena. In condensed matter physics it is well known that an effective Heisenberg Hamiltonian allows one to catch essential properties of the system, although the real structure cannot be expected to be reproduced in the details. One also should keep in mind that models like the mentioned Ginzburg-Landau effective theory of superconductivity perfectly models the phase transition between type I, type II superconductivity and normal state, without reference to the true microscopic structure. In any case it is always possible to find out to what extent the description fits to reality.
It is well known that long-range physics manifests itself as field theory naturally from underlying classical statistical systems exhibiting short-range exchange interactions (e.g. nearest-neighbor interactions on a lattice system) [111,112] (for lectures on the topic see e.g. [113]). However, the Planck system not only shows typical shortrange interactions. We know that it also features a long-range gravitational potential, which develops multipole excitations showing up as spin 1, spin 2 and higher modes at long distances [126].
In our context what is important is that the quadratic and quartic enhancements are persisting, as well as the running (screening or anti-screening effects) of couplings and their competition and conspiracy. Together these elements manifest themselves in the existence of the zeros of the enhanced terms, provided these zeros are not to close to the cutoff. A look at Fig. 7 shows that such effects can be dramatic fairly well below the cutoff. Again, the perturbativeness, together with the fact that leading corrections to these results are by dim 6 operators, let us expect that results are reliable at the 10 −4 level up to 10 17 GeV, which is in the middle of the symmetric phase already. Once the phase transition has happened, the running is anyway weak and even when cutoff-effect are starting to play a role they cannot spoil the relevant qualitative features concerning triggering inflation, reheating and related phenomena.
Lattice SM simulations in the appropriate parameter range of vacuum stability, keeping top-quark Yukawa and Higgs self-coupling to behave asymptotically free, which requires to include simultaneously besides the Higgs system also the top-quark Yukawa sector and QCD, could help to investigate such problems quantitatively. Experience from lattice QCD simulations may not directly be illustrative since usually the cutoff is rather close and a crucial difference is also the true non-perturbative nature of low energy QCD.
In any case, not to include the effects related to the relevant operators (dim < 4) simply must give wrong results. Even substantial uncertainties, which certainly show up closer to the cutoff, in power-like behaved quantities seem to be an acceptable shortcoming in comparison to not taking into account the cutoff-enhancements at all (as usually done).

Summary and Conclusions
A cutoff regularized SM with the Planck mass as a cutoff is considered to exhibit the relevant features of the physical Planck world in the sense that it resides in the same universality class with respect to its long-range behavior. The SM we observe at low energy is then the emergent renormalizable effective theory of the Planck medium. All conditions that usually have to be imposed as principles to ensure renormalizability are emergent as a result of the low energy expansion. The relation between long distance physics and short distance physics can be controlled in principle via Wilson's renormalization semi-group. Taking into account renormalization effects and the "running" of the parameters is mandatory in order to understand what has been happening in the evolution of the universe. The key outcome of the LEESM setting is that for appropriately tuned parameters the quadratic as well as the quartic "singularities" are nullified at a specific energy, where bosonic and fermionic effects cancel. 28 In the SM the corresponding zeros happen because the bosonic and the fermionic contributions are running differently, according to their respective renormalization groups, and at some point turn out to be of equal magnitude.
In this scenario the Higgs field/particle has two different functions in our world: (1) it has to render the effective low energy electroweak theory (massive vector-boson and fermion sector) renormalizable. In place of fermion mass terms we have fermion Yukawa couplings to start with, while gauge boson mass terms enter via the kinetic Higgs term through the covariant derivative that has to include the gauge fields. In the broken low energy phase the Higgs field forms a vacuum condensate, which pro-vides masses to all massive fields including the Higgs boson itself. Key point is the many new Higgs field exchange-forces necessary to render the low energy amplitudes renormalizable. (2) in the symmetric phase there exist four very heavy Higgs bosons (H , φ 0 , H ± ) that generate a huge positive dark energy, as required to triggers inflation. After inflation has ended and we are out of equilibrium the Higgs bosons are decaying predominantly into the fermions pairs with largest Yukawa couplings (predominantly at this stage still massless top-antitop pairs), which provides the reheating of the inflated universe. The universe cooling further down then pushes the universe into the Higgs phase, where the particles acquire their masses. The predominating heavy quarks decay into the lighter ones, which later form the baryons and normal matter. This scenario is possible because of the quadratically enhanced Higgs boson mass and the quartically enhanced dark energy, which show up in the symmetric phase of the SM before the transition into the Higgs phase. The existence of such relevant operator effects in my opinion is supported by observed inflation patterns. Both, the hierarchy "problem" as well as the cosmological constant "problem" reflect important properties of the SM needed to understand the evolution of the early universe (for different opinions see [174][175][176][177][178]). Consolidation of our bottom-up path to physics near the Planck scale will sensibly depend on progress in high precision physics around the EW scale v. Especially, Higgs boson and top-quark factories (like FCC-ee or ILC) will play a key role in this context.
Concerning the presumed fine-tuning problem: the scales M Pl and v stand for different regimes and there is no reason why they should not be vastly different; one is related to gravity (Planck medium) the other to long-range order at low energies. It is the energy dependence of the SM interactions that triggers spontaneous symmetry breaking. The emergent SM symmetry apparently only three orders of magnitude below M Pl gets broken by a non-symmetric ground state rearrangement. The critical point nevertheless is the actual value of v that is non-vanishing only below a critical temperature. While in a condensed matter system one is adjusting the temperature by hand, the key problem seems to be that in particle physics we cannot adjust the temperature. But this the expanding universe is doing for us. In fact, the hot big bang universe provides a scan of the temperature spectrum and automatically triggers the phase transition at some point, as the calculations show. Temporary out of equilibrium phases do not disturb the gross behavior of this scan, but they will have to be investigated in any case. For more details I refer to my original articles [1,2] and to my Krakow Lectures [179].
The scenario I advocate requires a change of paradigm, to one where the SM with its structure is emergent from a Planck cutoff-medium following a minimal selforganized "strategy", i.e. conspiracies are taking place to make structures to show up at large distances (for a short history of the development of the emergence paradigm see e.g. [180]). This looks like a version of an anthropic principle at work. It lets look the SM to be more natural than many of the BSM scenarios we have heard about during the last about 45 years. Although the SM started to turn out to work well and to work better than ever expected, we know it cannot explain a number of observationally established facts. Yes, the SM misses dark matter, singlet neutrinos, axions and likely more, but all of these have room in an emergence scenario. This is in contrast to the top-bottom philosophy behind the most popular BSM physics scenarios like string theory, supersymmetry or grand unification, which assume that the short distance world is intrinsically highly symmetric and symmetries are broken spontaneously only, because renormalizability is always assumed to be a basic law of nature. We know that the SM as seen at low energies is a spontaneously broken gauge theory, which gets more symmetric as we go to higher energies because mass operators in a high-energy expansion turn into irrelevant operators. This may have lead to a wrong perception concerning what we have to expect on the path to higher energies. This view overlooks the fact that a tower of possible symmetry breaking irrelevant operators of the low energy expansion turn into relevant contributions at high energies. Thus symmetries seen at low energy usually do not persist at higher energies. The dream that an eternal highly symmetric "theory of everything" should sit at and above the Planck scale may not be very realistic. The opposite very probably is true, the world looks more complex the closer we look, and symmetries emerge from not resolving the detailed structure behind. And the final truth remains something we can get closer only but will never be reached.
Between my Higgs inflation scenario and the metastability scenery favored in [65,68,85,86], the major difference is that for me understanding the relationship between the physical low energy parameters and the bare parameter, assumed to become the physical ones at short distances, is the mandatory premise. Most other analyses are working with a renormalized effective potential in the broken phase all the way up towards the Planck scale and do not consider cutoff effects to be physical. Power enhanced cutoff-effects in my scenario are triggering a phase transition between an early symmetric and a later broken phase. Higgs system conditioned inflation is possible only if in the early universe the SM is in the symmetric phase.
I think the LEESM scenario has a good chance to find its confirmation along the lines described in this article. Many aspects need to be checked and possibly modified. Admittedly, there are many open questions, which should be investigated more thoroughly. One conclusion seems to be unavoidable, namely that the SM Higgs sector provides dark energy that affects early as well as late cosmology. Obviously, discovering SUSY, a GUT, extra-dimensions, a little Higgs extension, Technicolor or similar extensions of the SM would spoil the scene. Notably, also one more fermion family or one less would completely mess up everything. The sharp dependence of the Higgs vacuum stability on the SM input parameters and on possible SM extensions and the vastly different scenarios which can result as a consequence of minor shifts in parameter space makes the stable vacuum case a particularly interesting one and it could reveal the Higgs particle as "the master of the universe". After all it is commonly accepted that dark energy is the"stuff" shaping the universe both at very early as well as at the late times.

Appendix: How Natural is the Minimal SM?
Often it is considered that it would be more natural to have a left-right symmetric world including mirror fermions. The following consideration, which goes back to Veltman [181] (see also [182][183][184][185]), is instructive as it helps to understand why parity violation is quite natural and why QED conserves parity. It has a lot to do with the assumption of a minimal Higgs system. I reproduce a version, which I had presented in [186] some time ago. Actually within the context of our LEESM scenario we gain a much deeper insight, because the assumptions made are now emergent properties resulting from the low energy expansion.
In order to try to derive the SM let us make the following assumptions: (1) local field theory (2) interactions follow from a local gauge principle (3) renormalizability (4) masses derive from the minimal Higgs system (5) ν R which we know must exist does not carry hypercharge. Note that all points besides the last one are emergent structures in a LEESM as we may learn from [126][127][128][129][130][131][132] (see also Sect. 3). We admit that the last assumption looks somewhat ad hoc, but nevertheless we make it. From the above assumptions the following picture develops: -For the gauge interactions the simplest non-trivial possibility is that the fundamental massless matter fields group according to the simplest possibilities, into doublets and triplets, which are the fundamental representations of SU (2) and SU (3), besides possible singlets. -Since fields are massless all fields can be chosen left-handed. Left-handed particles and left-handed antiparticles at this stage are uncorrelated. -We must have pairing for particles that are going to be massive, since a mass term (we ignore the possibility to have Majorana fields here) has the form ψψ =ψ L ψ R +ψ R ψ L . Notice that for massive particles only, we know which left-handed antiparticle belongs to which left-handed particle to form a Dirac field. -For SU(3) c triplets we must have pairing in order to avoid axial anomalies.
SU (3) is the simplest group having complex representations. This allows to put particles in 3 and antiparticles in the inequivalent 3 * . As a consequence a rich color singlet structure (≡ hadron spectrum) results. Furthermore, confinement requires SU(3) c to be unbroken ! -SU(2) L is anomaly free and hence there is no anomaly condition associated with this group. To generate mass we have to break SU(2) L by a Higgs mechanism. The simplest and natural possibility is to choose one Higgs field in the fundamental representation of SU(2) L . There is no hypercharge for the moment. The Higgs field may be written in the form in terms of a 2 × 2 matrix field (τ i , i = 1, 2, 3 the Pauli matrices) The covariant derivative being given by and the SU(2) invariant renormalizable Higgs system exhibits an extra global SU(2) R -symmetry χ b → V + χ b . One easily checks that the transformation with U (x) ∈ SU(2) L,local , V ∈ SU(2) R,global leaves the Higgs Lagrangian invariant. This implies that the fields (W + , W 3 , W − ) form a weak isospin triplet with M Z = M W ± . Now consider the fermions (still no hypercharge). Since L f and Φ b are SU (2) doublets there also must exist singlet fermions R f , otherwise we would not be able to write down an invariant and renormalizable fermion-Higgs coupling. Therefore SU(2) L must be parity violating of V-A-type! The Yukawa term has the general form L Yukawa = −L f Φ g 11 g 12 g 21 g 22 R f + h.c. , with 4 complex couplings g i j and R f a "doublet" having two right-handed singlets as entries. Although we have not used hypercharge to restrict these couplings the existence of a global SU(2) R -symmetry of the Higgs system allows us to transform the Yukawa couplings Φ (·) R f → ΦV + (·)W R f to standard form, V + (·)W = real diagonal. Since V ∈ SU(2) R has 3 parameters and W is an arbitrary unitary matrix with 4 parameters we end up with one free parameter such that the system exhibits a global U (1) invariance. This is not surprising since in the unitary gauge we always can end up only with L Yukawa in the simple standard form -The global U (1) which is a consequence of the minimal Higgs mechanism may be interpreted as a global U (1) Y . We are free to assign Y = 1 to Φ b , which means nothing else than that we measure Y in units of the Φ b -hypercharge. Then Φ t = Φ χ t ; χ t = 1 0 has Y = −1 , and we may write Φ = (Φ b , Φ t ). Since we have the global U (1) Y for free, we may assume this symmetry to be local. The covariant derivative for Φ now reads and we find back the usual Higgs Lagrangian The 3 real fields φ a a = 1, 2, 3 could and have been gauged away and only 3 out of 4 gauge fields can acquire a mass. Hence there must exist one massless field, the photon! Evidently we obtain the relations g = g tan Θ W and ρ = where Y L is arbitrary. By inspection of L Yukawa we find for the hypercharges of the singlets: ψ 1R must have Y 1R = Y L + 1 and ψ 2R must have Y 2R = Y L − 1.
One consequence is that U (1) Y must violate parity. The astonishing thing is that the fermion current which couples to the photon preserves parity. By inspection we find and the couplings of L f and R f to A μ read Because we have g cos Θ W = g sin Θ W = e we find the Gell-Mann-Nishijima relation as a consequence of a minimal Higgs structure! What we find is that, whatever the hypercharge of L f is, L f and R f must couple identically to photons. Thus QED must be parity conserving! Furthermore, the charges of the upper (1) and lower (2) components of the doublets satisfy Q Li = Q Ri , Q 1 − Q 2 = 1 and Q 1 + Q 2 = Y L .
So far we have no charge quantization. Here we need the last assumption. -If ν R does not couple to the U (1) gauge field, we have to set Y ν R = 0 and consequently we must have Y ν L = −1 = Y L = 0 and Q ν = 0, Q = −1. For the U (1) Y anomaly cancellation we need lepton-quark duality and the charges of the quarks must have their known values if they appear in three colors. One thus must have the usual charge quantization.
We finally summarize the consequences of the assumptions stated above: -Breaking SU(2) L by a minimal Higgs automatically leads to a global U (1) Y , which can be gauged -parity violation of SU(2) L ρ = M 2 W /(M 2 Z cos 2 Θ W ) = 1 -the existence of the photon -parity conservation of QED -the validity of the Gell-Mann-Nishijima relation -family structure -charge quantization.
We do not know why right-handed neutrinos are sterile i.e. do not couple to gauge bosons. In the SM of electroweak interactions, neutrinos originally were assumed to be massless i.e. that right-handed neutrinos did not exist. This is definitely ruled out by the observation of neutrino oscillations. 29 I think this reasoning is able to help understanding how various excitations in the chaotic Planck medium develop a pattern like the SM as a low energy effective 29 We note that right-handed neutrinos have been expected to exist from the point of view of the SM gauge symmetry. They have been excluded by imposing additional global flavor conservation, because the latter had not been observed within the achieved experimental accuracy. In fact flavor violation is so tiny that a direct observation still is missing. The corresponding smallness of neutrino masses very likely is a consequence of a see-saw mechanism, which could be triggered within the SM when the singlets would be of Majorana type. Majorana mass terms are not protected by any symmetry and therefore would be expected to be related to the Planck scale in a similar way as the Higgs field mass in the symmetric phase. structure. Renormalizability as a consequence of the low energy expansion and the very large gap between the EW and the Planck scales plus a certain minimality (not too little but not too much e.g. only up to symmetry triplets) determines the SM structure without much freedom. After all, minimality is not a new concept in physic as we know from the principle of least action. Three fermion families are required in order CP violation emerges in a natural way, and to make baryogenesis eventually possible within the LEESM scenario as addressed in Sect. 7. We have been emphasizing the high self-consistency of the SM where all essential structures look to be emergent properties in the low energy effective viewport of a cutoff-system residing at the Planck scale. "What is not capable of surviving at long distances does not exist there" (Darwin revisited).