tag:blogger.com,1999:blog-86694682016-09-21T14:47:23.796+01:00Life on the latticeThoughts on lattice QCD, particle physics and the world at large.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.comBlogger214125tag:blogger.com,1999:blog-8669468.post-35531267446787708822016-07-30T21:50:00.002+01:002016-07-30T21:51:40.802+01:00Lattice 2016, Day SixThe final day of the conference started with a review talk by Claudio Pica on lattice simulations trying to chart the fundamental physics beyond the Standard Model. The problem with the SM is perhaps to some extent how well it works, given that we know it must be incomplete. One of the main contenders for replacing it is the notion of strong dynamics at a higher energy scale giving rise to the Higgs boson as a composite particle. The most basic "technicolor" theories of this kind fail because they cannot account for the relatively large masses of the second- and third-generation quarks. To avoid that problem, the coupling of the technicolor gauge theory must not be running, but "walking" slowly from high to low energy scales, which has given rise to a veritable industry of lattice simulations investigating the β function of various gauge theories coupled to various numbers of fermions in various representations. The Higgs can then be either a dilaton associated with the breaking of conformal symmetry, which would naturally couple like a Standard Model Higgs, or a pseudo-Goldstone boson associated with the breaking of some global flavour symmetry. So far, nothing very conclusive has resulted, but of course the input from experiment at the moment only consists of limits ruling some models out, but not allowing for any discrimination between those models that aren't rules out.<br /><br />A specific example of BSM physics, <i>viz.</i> strongly interacting dark matter, was presented in a talk by Enrico Rinaldi. If there is a new strongly-coupled interaction, as suggested by the composite Higgs models, then besides the Higgs there will also be other bound states, some of which may be stable and provide a dark matter candidate. While the "dark" nature of dark matter requires such a bound state to be neutral, the constituents might interact with the SM sector, allowing for the production and detection of dark matter. Many different models of composite dark matter have been considered, and the main limits currently come from the non-detection of dark matter in searches, which put limits on the "hadron-structure" observables of the dark matter candidates, such as their σ-terms and charge radii).<br /><br />David Kaplan gave a talk on a new perspective on chiral gauge theories, the lattice formulation of which has always been a persistent problem, largely due to the Nielsen-Ninomiya theorem. However, the fermion determinant of chiral gauge theories is already somewhat ill-defined even in the continuum. A way to make it well-defined has been proposed by Alvarez-Gaumé <i>et al.</i> through the addition of an ungauged right-handed fermion. On the lattice, the U(1)<sub>A</sub> anomaly is found to emerge as the remnant of the explicit breaking of chiral symmetry by e.g. the Wilson term in the limit of vanishing lattice spacing. Attempts at realizing ungauged mirror fermions using domain wall fermions with a gauge field constrained to near one domain wall have failed, and a realizations using the gradient flow in the fifth dimension turns the mirror fermions into "fluff". A new realization along the lines of the overlap operator gives a lattice operator very similar to that of Alvarez-Gaumé by coupling the mirror fermion to a fixed point of the gradient flow, which is a pure gauge.<br /><br />After the coffee break, Tony Hey gave a very entertaining, if somewhat meandering, talk about "Richard Feynman, Data-Intensive Science and the Future of Computing" going all the way from Feynman's experiences at Los Alamos to AI singularity scenarios and the security aspects of self-driving cars.<br /><br />The final plenary talk was the review talk on machines and algorithms by Peter Boyle. The immediate roadmap for new computer architectures shows increases of around 400 times in the single-precision performance per node, and a two-fold increase in the bandwidth of interconnects, and this must be taken into account in algorithm design and implementation in order to achieve good scaling behaviour. Large increases in chip performance are to be expected from three-dimensional arrangement of units, which will allow thicker and shorter copper wires, although there remain engineering problems to solve, such as how to efficiently get the heat out of such chips. In terms of algorithms, multigrid solvers are now becoming available for a larger variety of fermion formulations, leading to potentially great increases in performance near the chiral and continuum limits. Multilevel integration methods, which allow for an exponential reduction of the noise, also look interesting, although at the moment these work only in the quenched theory.<br /><br />The IAC announced that Lattice 2018 will take place at Michigan State University. Elvira Gamiz as the chair of the Lattice 2017 LOC extended an invitation to the lattice community to come to Granada for <a href="http://www.lattice2017.es/">Lattice 2017</a>, which will take place in the week 18-24 June 2017. And with that, and a round of well-deserved applause for the organizers, the conference closed.<br /><br />My further travel plans are of interest only to a small subset of my readers, and need not be further elaborated upon in this venue.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-13123404377088694202016-07-29T22:26:00.003+01:002016-07-29T22:26:20.682+01:00Lattice 2016, Day FiveToday was the day of finite temperature and density, on which the general review talk was delivered by Heng-Tong Ding. While in the meantime agreement has been reached on the transition temperature, the nature of the transition (crossover) and the equation of state at the physical quark masses, on which different formulations differed a lot in the past, the Columbia plot of the nature of the transition as a function of the light and strange quark masses still remains to be explored, and there are discrepancies between results obtained in different formulations. On the topic of U(1)<sub>A</sub> restoration (on which I do have a layman's question: to my understanding U(1)<sub>A</sub> is broken by the axial anomaly, which to my understanding arises from the path integral measure - so why should one expect the symmetry to be restored at high temperature? The situation is quite different from dynamical spontaneous symmetry breaking, as far as I understand), there is no evidence for restoration so far. A number of groups have taken to using the gradient flow as a tool to perform relatively cheap investigations of the equation of state. There are also new results from the different approaches to finite-density QCD, including cumulants from the Taylor-expansion approach, which can be related to heavy-ion observables, and new ways of stabilizing complex Langevin dynamics.<br /><br />This was followed by two topical talks. The first, by Seyong Kim, was on the subject of heavy flavours at finite temperature. Heavy flavours are one of the most important probes of the quark-gluon plasma, and J/ψ suppression has served as a diagnostic tool of QGP formation for a long time. To understand the influence of high temperatures on the survival of quarkonium states and on the transport properties of heavy flavours in the QGP, knowledge of the spectral functions is needed. Unfortunately, extracting these from a finite number of points in Euclidean point is an ill-posed problem, especially so when the time extent is small at high temperature. The methods used to get at them nevertheless, such as the maximum entropy method or Bayesian fits, need to use some kind of prior information, introducing the risk of a methodological bias leading to systematic errors that may be not only quantitative, but even qualitative; as an example, MEM shows P-wave bottomonium to melt around the transition temperature, whereas a newer Bayesian method shows it to survive, so clearly more work is needed.<br /><br />The second topical talk was Kurt Langfeld speaking about the density-of-states method. This method is based on determining a function ρ(E), which is essentially the path integral of δ(S[φ]-E), such that the partition function can be written as the Laplace transform of ρ, which can be generalized to the case of actions with a sign problem, where the partition function can then be written as the Fourier transform of a function P(s). An algorithm to compute such functions exists in the form of what looks like a sort of microcanonical simulation in a window [E-δE;E+δE] and determines the slope of ρ at E, whence ρ can be reconstructed. Ergodicity is ensured by having the different windows overlap and running in parallel, with a possibility of "replica exchange" between the processes running for neighbouring windows when configurations within the overlap between them are generated. The examples shown, e.g. for the Potts model, looked quite impressive in that the method appears able to resolve double-peak structures even when the trough between the peaks is suppressed by many orders of magnitude, such that a Markov process would have no chance of crossing between the two probability peaks.<br /><br />After the coffee break, Aleksi Kurkela reviewed the phenomenology of heavy ions. The flow properties that were originally taken as a sign of hydrodynamics having set in are now also observed in pp collisions, which seem unlikely to be hydrodynamical. In understanding and interpreting these results, the pre-equilibration evolution is an important source of uncertainty; the current understanding seems to be that the system goes from an overoccupied to an underoccupied state before thermalizing, making different descriptions necessary at different times. At early times, simulations of classical Yang-Mills theory on a lattice in proper-time/rapidity coordinates are used, whereas later a quasiparticle description and kinetic theory can be applied; all this seems to be qualitative so far.<br /><br />The energy momentum tensor, which plays an important role in thermodynamics and hydrodynamics, was the topic of the last plenary of the day, which was given by Hiroshi Suzuki. Translation invariance is broken on the lattice, so the Ward-Takahashi identity for the energy-momentum tensor picks up an O(a) violation term, which can become O(1) by radiative corrections. As a consequence, three different renormalization factors are needed to renormalize the energy-momentum tensor. One way of getting at these are the shifted boundary conditions of Giusti and Meyer, another is the use of the gradient flow at short flow times, and there are first results from both methods.<br /><br />The parallel sessions of the afternoon concluded the parallel programme.<br /><br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-73996963967664543032016-07-28T23:59:00.000+01:002016-07-29T17:37:26.483+01:00Lattice 2016, Days Three and FourFollowing the canonical script for lattice conferences, yesterday was the day without plenaries. Instead, the morning was dedicated to parallel sessions (including my own talk), and the afternoon was free time with the option of taking one of several arranged excursions.<br /><br />I went on the excursion to Salisbury cathedral (which is notable both for its fairly homogeneous and massive architectural ensemble, and for being home to one of four original copies of the Magna Carta) and Stonehenge (which in terms of diameter seems to be much smaller than I had expected from photos).<br /><br />Today began with the traditional non-lattice theory talk, which was given by Monika Blanke, who spoke about the impact of lattice QCD results on CKM phenomenology. Since quarks cannot be observed in isolation, the extraction of CKM matrix elements from experimental results always require knowledge of the appropriate hadronic matrix elements of the currents involved in the measured reaction. This means that lattice results for the form factors of heavy-to-light semileptonic decays and for the hadronic parameters governing neutral kaon and B meson mixing are of crucial importance to CKM phenomenology, to the extent that there is even a sort of "wish list" to the lattice. There has long been a discrepancy between the values of both |V<sub>cb</sub>| and |V<sub>ub</sub>| extracted from inclusive and exclusive decays, respectively, and the ratio |V<sub>ub</sub>/V<sub>cb</sub>| that can be extracted from decays of Λ<sub>b</sub> baryons only adds to the tension. However, this is likely to be a result of underestimated theoretical uncertainties or experimental issues, since the pattern of the discrepancies is not in agreement with that which would results from new physics effects induced by right-handed currents. General models of flavour violating new physics seems to favour the inclusive value for |V<sub>ub</sub>|. In b->s transitions, there is evidence for new physics effects at the 4σ level, but significant theoretical uncertainties remain. The B<sub>(s)</sub>->μ<sup>+</sup>μ<sup>-</sup> branching fractions are currently in agreement with the SM at the 2σ level, but new, more precise measurements are forthcoming.<br /><br />Ran Zhou complemented this with a review talk about heavy flavour results from the lattice, where there are new results from a variety of different approaches (NRQCD, HQET, Fermilab and Columbia RHQ formalisms), which can serve as useful and important cross-checks on each other's methodological uncertainties.<br /><br />Next came a talk by Amy Nicholson on neutrinoless double β decay results from the lattice. Neutrinoless double β decays are possible if neutrinos are Majorana particles, which would help to explain the small masses of the observed left-handed neutrinos through the see-saw mechanism pushing the right-handed neutrinos off to near the GUT scale. Treating the double β decay in the framework of a chiral effective theory, the leading-order matrix element required is a process π<sup>-</sup>->π<sup>+</sup>e<sup>-</sup>e<sup>-</sup>, for which there are first results in lattice QCD. The NLO process would have disconnected diagrams, but cannot contribute to the 0<sup>+</sup>->0<sup>+</sup> transitions which are experimentally studied, whereas the NNLO process involves two-nucleon operators and still remains to be studied in greater detail on the lattice.<br /><br />After the coffee break, Agostino Patella reviewed the hot topic of QED corrections to hadronic observables. There are currently two main methods for dealing with QED in the context of lattice simulations: either to simulate QCD+QED directly (usually at unphysically large electromagnetic couplings followed by an extrapolation to the physical value of α=1/137), or to expand it in powers of α and to measure only the resulting correlation functions (which will be four-point functions or higher) in lattice QCD. Both approaches have been used to obtain some already very impressive results on isospin-breaking QED effects in the hadronic spectrum, as shown already in the spectroscopy review talk. There are, however, still a number of theoretical issues connected to the regularization of IR modes that relate to the Gauss law constraint that would forbid the existence of a single charged particle (such as a proton) in a periodic box. The prescriptions to evade this problem all lead to a non-commutativity of limits requiring the infinite-volume limit to be taken before other limits (such as the continuum or chiral limits): QED<sub>TL</sub>, which omits the global zero modes of the photon field, is non-local and does not have a transfer matrix; QED<sub>L</sub>, which omits the spatial zero modes on each timeslice, has a transfer matrix, but is still non-local and renormalizes in a non-standard fashion, such that it does not have a non-relativistic limit; the use of a massive photon leads to a local theory with softly broken gauge symmetry, but still requires the infinite-volume limit to be taken before removing the photon mass. Going beyond hadron masses to decays introduces new IR problems, which need to be treated in the Bloch-Nordsieck way, leading to potentially large logarithms.<br /><br />The 2016 Ken Wilson Lattice Award was awarded to Antonin Portelli for his outstanding contributions to our understanding of electromagnetic effects on hadron properties. Antonin was one of the driving forces behind the BMW collaboration's effort to determine the proton-neutron mass difference, which resulted in a <i>Science</i> paper exhibiting one of the most frequently-shown and impressive spectrum plots at this conference.<br /><br />In the afternoon, parallel sessions took place, and in the evening there was a (very nice) conference dinner at the Southampton F.C. football stadium.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-77335955127742245402016-07-26T21:32:00.001+01:002016-07-26T21:32:12.667+01:00Lattice 2016, Day TwoHello again from Lattice 2016 at Southampton. Today's first plenary talk was the review of nuclear physics from the lattice given by Martin Savage. Doing nuclear physics from first principles in QCD is obviously very hard, but also necessary in order to truly understand nuclei in theoretical terms. Examples of needed theory predictions include the equation of state of dense nuclear matter, which is important for understanding neutron stars, and the nuclear matrix elements required to interpret future searches for neutrinoless double β decays in terms of fundamental quantities. The problems include the huge number of required quark-line contractions and the exponentially decaying signal-to-noise ratio, but there are theoretical advances that increasingly allow to bring these under control. The main competing procedures are more or less direct applications of the Lüscher method to multi-baryon systems, and the HALQCD method of computing a nuclear potential from Bethe-Salpeter amplitudes and solving the Schrödinger equation for that potential. There has been a lot of progress in this field, and there are now first results for nuclear reaction rates.<br /><br />Next, Mike Endres spoke about new simulation strategies for lattice QCD. One of the major problems in going to very fine lattice spacings is the well-known phenomenon critical slowing-down, i.e. the divergence of the autocorrelation times with some negative power of the lattice spacing, which is particularly severe for the topological charge (a quantity that cannot change at all in the continuum limit), leading to the phenomenon of "topology freezing" in simulations at fine lattice spacings. To overcome this problem, changes in the boundary conditions have been proposed: open boundary conditions that allow topological charge to move into and out of the system, and non-orientable boundary conditions that destroy the notion of an integer topological charge. An alternative route lies in algorithmic modifications such as metadynamics, where a potential bias is introduced to disfavour revisiting configurations, so as to forcibly sample across the potential wells of different topological sectors over time, or multiscale thermalization, where a Markov chain is first run at a coarse lattice spacing to obtain well-decorrelated configurations, and then each of those is subjected to a refining operation to obtain a (non-thermalized) gauge configuration at half the lattice spacing, each of which can then hopefully thermalized by a short sequence of Monte Carlo update operations.<br /><br />As another example of new algorithmic ideas, Shinji Takeda presented tensor networks, which are mathematical objects that assign a tensor to each site of a lattice, with lattice links denoting the contraction of tensor indices. An example is given by the rewriting of the partition function of the Ising model that is at the heart of the high-temperature expansion, where the sum over the spin variables is exchanged against a sum over link variables taking values of 0 or 1. One of the applications of tensor networks in field theory is that they allow for an implementation of the renormalization group based on performing a tensor decomposition along the lines of a singular value decomposition, which can be truncated, and contracting the resulting approximate tensor decomposition into new tensors living on a coarser grid. Iterating this procedure until only one lattice site remains allows the evaluation of partition functions without running into any sign problems and at only O(log <i>V</i>) effort.<br /><br />After the coffee break, Sara Collins gave the review talk on hadron structure. This is also a field in which a lot of progress has been made recently, with most of the sources of systematic error either under control (e.g. by performing simulations at or near the physical pion mass) or at least well understood (e.g. excited-state and finite-volume effects). The isovector axial charge <i>g<sub>A</sub></i> of the nucleon, which for a long time was a bit of an embarrassment to lattice practitioners, since it stubbornly refused to approach its experimental value, is now understood to be particularly severely affected by excited-state effects, and once these are well enough suppressed or properly accounted for, the situation now looks quite promising. This lends much larger credibility to lattice predictions for the scalar and tensor nucleon charges, for which little or no experimental data exists. The electromagnetic form factors are also in much better shape than one or two years ago, with the electric Sachs form factor coming out close to experiment (but still with insufficient precision to resolve the conflict between the experimental electron-proton scattering and muonic hydrogen results), while now the magnetic Sachs form factor shows a trend to undershoot experiment. Going beyond isovector quantities (in which disconnected diagrams cancel), the progress in simulation techniques for disconnected diagrams has enabled the first computation of the purely disconnected strangeness form factors. The sigma term σ<sub>πN</sub> comes out smaller on the lattice than it does in experiment, which still needs investigation, and the average momentum fraction <<i>x</i>> still needs to become the subject of a similar effort as the nucleon charges have received.<br /><br />In keeping with the pattern of having large review talks immediately followed by a related topical talk, Huey-Wen Lin was next with a talk on the Bjorken-<i>x</i> dependence of the parton distribution functions (PDFs). While the PDFs are defined on the lightcone, which is not readily accessible on the lattice, a large-momentum effective theory formulation allows to obtain them as the infinite-momentum limit of finite-momentum parton distribution amplitudes. First studies show interesting results, but renormalization still remains to be performed.<br /><br />After lunch, there were parallel sessions, of which I attended the ones into which most of the <i>(g-2)</i> talks had been collected, showing quite a rate of progress in terms of the treatment of in particular the disconnected contributions.<br /><br />In the evening, the poster session took place.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-67540110959609539562016-07-25T22:58:00.001+01:002016-07-25T23:05:04.594+01:00Lattice 2016, Day OneHello from Southampton, where I am attending the Lattice 2016 conference.<br /><br />I arrived yesterday safe and sound, but unfortunately too late to attend the welcome reception. Today started off early and quite well with a full English breakfast, however.<br /><br />The conference programme was opened with a short address by the university's Vicepresident of Research, who made a point of pointing out that he like 93% of UK scientists had voted to remain in the EU - an interesting testimony to the political state of affairs, I think.<br /><br />The first plenary talk of the conference was a memorial to the scientific legacy of Peter Hasenfratz, who died earlier this year, delivered by Urs Wenger. Peter Hasenfratz was one of the pioneers of lattice field theory, and hearing of his groundbreaking achievements is one of those increasingly rare occasions when I get to feel very young: when he organized the first lattice symposium in 1982, he sent out individual hand-written invitations, and the early lattice reviews he wrote were composed in a time where most results were obtained in the quenched approximation. But his achievements are still very much current, amongst other things in the form of fixed-point actions as a realization of the Ginsparg-Wilson relation, which gave rise to the booming interest in chiral fermions.<br /><br />This was followed by the review of hadron spectroscopy by Chuan Liu. The contents of the spectroscopy talks have by now shifted away from the ground-state spectrum of stable hadrons, the calculation of which has become more of a benchmark task, and towards more complex issues, such as the proton-neutron mass difference (which requires the treatment of isospin breaking effects both from QED and from the difference in bare mass of the up and down quarks) or the spectrum of resonances (which requires a thorough study of the volume dependence of excited-state energy levels via the Lüscher formalism). The former is required as part to the physics answer to the ageless question why anything exists at all, and the latter is called for in particular by the still pressing current question of the nature of the XYZ states.<br /><br />Next came a talk by David Wilson on a more specific spectroscopy topic, namely resonances in coupled-channel scattering. Getting these right requires not only extensions of the Lüscher formalism, but also the extraction of very large numbers of energy levels via the generalized eigenvalue problem.<br /><br />After the coffee break, Hartmut Wittig reviewed the lattice efforts at determining the hadronic contributions to the anomalous magnetic moment (g-2)<sub>μ</sub> of the muon from first principles. This is a very topical problem, as the next generation of muon experiments will reduce the experimental error by a factor of four or more, which will require a correspondingly large reduction in the theoretical uncertainties in order to interpret the experimental results. Getting to this level of accuracy requires getting the hadronic vacuum polarization contribution to sub-percent accuracy (which requires full control of both finite-volume and cut-off effects, and a reasonably accurate estimate for the disconnected contributions) and the hadronic light-by-light scattering contribution to an accuracy of better than 10% (which some way or another requires the calculation of a four-point function including a reasonable estimate for the disconnected contributions). There has been good progress towards both of these goals from a number of different collaborations, and the generally good overall agreement between results obtained using widely different formulations bodes well for the overall reliability of the lattice results, but there are still many obstacles to overcome.<br /><br />The last plenary talk of the day was given by Sergei Dubovsky, who spoke about efforts to derive a theory of the QCD string. As with most stringy talks, I have to confess to being far too ignorant to give a good summary; what I took home is that there is some kind of string worldsheet theory with Goldstone bosons that can be used to describe the spectrum of large-N<sub>c</sub> gauge theory, and that there are a number of theoretical surprises there.<br /><br />Since the plenary programme is being <a href="http://www.southampton.ac.uk/lattice2016/plenary-streaming/">streamed</a> on the web, by the way, even those of you who cannot attend the conference can now do without my no doubt quite biased and very limited summaries and hear and see the talks for yourselves.<br /><br />After lunch, parallel sessions took place. I found the sequence of talks by Stefan Sint, Alberto Ramos and Rainer Sommer about a precise determination of α<sub>s</sub>(M<sub>Z</sub>) using the Schrödinger functional and the gradient-flow coupling very interesting.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-21047487361222908442015-09-15T12:56:00.003+01:002015-10-03T20:13:12.118+01:00Fundamental Parameters from Lattice QCD, Last DaysThe last few days of our scientific programme were quite busy for me, since I had agreed to give the summary talk on the final day. I therefore did not get around to blogging, and will keep this much-delayed summary rather short.<br /><br />On Wednesday, we had a talk by Michele Della Morte on non-perturbatively matched HQET on the lattice and its use to extract the b quark mass, and a talk by Jeremy Green on the lattice measurement of the nucleon strange electromagnetic form factors (which are purely disconnected quantities).<br /><br />On Thursday, Sara Collins gave a review of heavy-light hadron spectra and decays, and Mike Creutz presented arguments for why the question of whether the up-quark is massless is scheme dependent (because the sum and difference of the light quark masses are protected by symmetries, but will in general renormalize differently).<br /><br />On Friday, I gave the summary of the programme. The main themes that I identified were the question of how to estimate systematic errors, and how to treat them in averaging procedures, the issues of isospin breaking and scale setting ambiguities as major obstacles on the way to sub-percent overall precision, and the need for improved communication between the "producers" and "consumers" of lattice results. In the closing discussion, the point was raised that for groups like CKMfitter and UTfit the correlations between different lattice quantities are very important, and that lattice collaborations should provide the covariance matrices of the final results for different observables that they publish wherever possible.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-59728605614272195742015-09-09T21:00:00.000+01:002015-09-15T12:55:40.616+01:00Fundamental Parameters from Lattice QCD, Day SevenToday's programme featured two talks about the interplay between the strong and the electroweak interactions. The first speaker was Gregorio Herdoíza, who reviewed the determination of hadronic corrections to electroweak observables. In essence these determinations are all very similar to the determination of the leading hadronic correction to (g-2)<sub>μ</sub> since they involve the lattice calculation of the hadronic vacuum polarisation. In the case of the electromagnetic coupling α, its low-energy value is known to a precision of 0.3 ppb, but the value of α(m<sub>Z</sub><sup>2</sup>) is known only to 0.1 ‰, and a larger portion of the difference in uncertainty is due to the hadronic contribution to the running of α, i.e. the hadronic vacuum polarization. Phenomenologically this can be estimated through the R-ratio, but this results in relatively large errors at low Q<sup>2</sup>. On the lattice, the hadronic vacuum polarization can be measured through the correlator of vector currents, and currently a determination of the running of α in agreement with phenomenology and with similar errors can be achieved, so that in the future lattice results are likely to take the lead here. In the case of the electroweak mixing angle, sin<sup>2</sup>θ<sub>w</sub> is known well at the Z pole, but only poorly at low energy, although a number of experiments (including the P2 experiment at Mainz) are aiming to reduce the uncertainty at lower energies. Again, the running can be determined from the Z-γ mixing through the associated current-current correlator, and current efforts are under way, including an estimation of the systematic error caused by the omission of quark-disconnected diagrams.<br /><br />The second speaker was Vittorio Lubicz, who looked at the opposite problem, i.e. the electroweak corrections to hadronic observables. Since approximately α=1/137, electromagnetic corrections at the one-loop level will become important once the 1% level of precision is being aimed for, and since the up and down quarks have different electrical charges, this is an isospin-breaking effect which also necessitates at the same time considering the strong isospin breaking caused by the difference in the up and down quark masses. There are two main methods to include QED effects into lattice simulations; the first is direct simulations of QCD+QED, and the second is the method of incorporating isospin-breaking effects in a systematic expansion pioneered by Vittorio and colleagues in Rome. Either method requires a systematic treatment of the IR divergences arising from the lack of a mass gap in QED. In the Rome approach this is done through splitting the Bloch-Nordsieck treatment of IR divergences and soft bremsstrahlung into two pieces, whose large-volume limits can be taken separately. There are many other technical issues to be dealt with, but first physical results from this method should be forthcoming soon.<br /><br />In the afternoon there was a discussion about QED effects and the range of approaches used to treat them.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-26208387252768829622015-09-07T18:00:00.000+01:002015-09-07T18:00:02.460+01:00Fundamental Parameters from Lattice QCD, Day SixThe second week of our Scientific Programme started with an influx of new participants.<br /><br />The first speaker of the day was Chris Kelly, who spoke about CP violation in the kaon sector from lattice QCD. As I hardly need to tell my readers, there are two sources of CP violation in the kaon system, the indirect CP-violation from neutral kaon-antikaon mixing, and the direct CP-violation from K->ππ decays. Both, however, ultimately stem from the single source of CP violation in the Standard Model, i.e. the complex phase e<sup>iδ</sup> in the CKM matrix, which gives the area of the unitarity triangle. The hadronic parameter relevant to indirect CP-violation is the kaon bag parameter B<sub>K</sub>, which is a "gold-plated" quantity that can be very well determined on the lattice; however, the error on the CP violation parameter ε<sub>K</sub> constraining the upper vertex of the unitarity triangle is dominated by the uncertainty on the CKM matrix element V<sub>cb</sub>. Direct CP-violation is particularly sensitive to possible BSM effects, and is therefore of particular interest. Chris presented the recent efforts of the RBC/UKQCD collaboration to address the extraction of the relevant parameter ε'/ε and associated phenomena such as the ΔI=1/2 rule. For the two amplitudes A<sub>0</sub> and A<sub>2</sub>, different tricks and methods were required; in particular for the isospin-zero channel, all-to-all propagators are needed. The overall errors are still large: although the systematics are dominated by the perturbative matching to the MSbar scheme, the statistical errors are very sizable, so that the 2.1σ tension with experiment observed is not particularly exciting or disturbing yet.<br /><br />The second speaker of the morning was Gunnar Bali, who spoke about the topic of renormalons. It is well known that the perturbative series for quantum field theories are in fact divergent asymptotic series, whose typical term will grow like <i>n<sup>k</sup>z<sup>n</sup>n!</i> for large orders <i>n</i>. Using the Borel transform, such series can be resummed, provided that there are no poles (IR renormalons) of the Borel transform on the positive real axis. In QCD, such poles arise from IR divergences in diagrams with chains of bubbles inserted into gluon lines, as well as from instanton-antiinstanton configurations in the path integral. The latter can be removed to infinity by considering the large-<i>N<sub>c</sub></i> limit, but the former are there to stay, making perturbatively defined quantities ambiguous at higher orders. A relevant example are heavy quark masses, where the different definitions (pole mass, MSbar mass, 1S mass, ...) are related by perturbative conversion factors; in a heavy-quark expansion, the mass of a heavy-light meson can be written as <i>M=m+Λ+O(1/m)</i>, where <i>m</i> is the heavy quark mass, and Λ a binding energy of the order of some QCD energy scale. As <i>M</i> is unambiguous, the ambiguities in <i>m</i> must correspond to ambiguities in the binding energy Λ, which can be computed to high orders in numerical stochastic perturbation theory (NSPT). After dealing with some complications arising from the fact that IR divergences cannot be probed directly in a finite volume, it is found that the minimum term in the perturbative series (which corresponds to the perturbative ambiguity) is of order 180 MeV in the quenched theory, meaning that heavy quark masses are only defined up to this accuracy. Another example is the gluon condensate (which may be of relevance to the extraction of α<sub>s</sub> from τ decays), where it is found that the ambiguity is of the same size as the typically quoted result, making the usefulness of this quantity doubtful.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-74189136302004808152015-09-04T19:00:00.000+01:002015-09-07T13:18:03.740+01:00Fundamental Parameters from Lattice QCD, Day FiveThe first speaker today was Martin Lüscher, who spoke about revisiting numerical stochastic perturbation theory. The idea behind numerical stochastic perturbation theory is to perform a simulation of a quantum field theory using the Langevin algorithm and to perturbatively expand the fields, which leads to a tower of coupled evolution equations, where only the lowest-order one depends explicitly on the noise, whereas the higher-order ones describe the evolution of the higher-order coefficients as a function of the lower-order ones. In Numerical Stochastic Perturbation Theory (NSPT), the resulting equations are integrated numerically (up to some, possibly rather high, finite order in the coupling), and the average over noises is replaced by a time average. The problems with this approach are that the autocorrelation time diverges as the inverse square of the lattice spacing, and that the extrapolation in the Langevin time step size is difficult to control well. An alternative approach is given by Instantaneous Stochastic Perturbation Theory (ISPT), in which the Langevin time evolution is replaced by the introduction of Gaussian noise sources at the vertices of tree diagrams describing the construction of the perturbative coefficients of the lattice fields. Since there is no free lunch, this approach suffers from power-law divergent statistical errors in the continuum limit, which arise from the way in which power-law divergences that cancel in the mean are shifted around between different orders when computing variances. This does not happen in the Langevin-based approach, because the Langevin theory is renormalizable.<br /><br />The second speaker of the morning was Siegfried Bethke of the Particle Data Group, who allowed us a glimpse at the (still preliminary) world average of α<sub>s</sub> for 2015. In 2013, there were five classes of α<sub>s</sub> determinations: from lattice QCD, τ decays, deep inelastic scattering, e<sup>+</sup>e<sup>-</sup> colliders, and global Z pole fits. Except for the lattice determinations (and the Z pole fits, where there was only one number), these were each preaveraged using the range method -- i.e. taking the mean of the highest and lowest central value as average, and assigning it an ncertainty of half the difference between them. The lattice results were averaged using a χ<sup>2</sup> weighted average. The total average (again a weighted average) was dominated by the lattice results, which in turn were dominated by the latest HPQCD result. For 2015, there have been a number of updates to most of the classes, and there is now a new class of α<sub>s</sub> determinations from the LHC (of which there is currently only one published, which lies rather low compared to other determinations, and is likely a downward fluctuation). In most cases, the new determinations have not or hardly changed the values and errors of their class. The most significant change is in the field of lattice determinations, where the PDG will change its policy and will no longer perform its own preaverages, taking instead the FLAG average as the lattice result. As a result, the error on the PDG value will increase; its value will also shift down a little, mostly due to the new LHC value.<br /><br />The afternoon discussion centered on α<sub>s</sub>. Roger Horsley gave an overview of the methods used to determine it on the lattice (ghost vertices, the Schrödinger functional, the static energy at short distances, current-current correlators, and small Wilson loops) and reviewed the criteria used by FLAG to assess the quality of a given determination, as well as the averaging procedure used (which uses a more conservative error than what a weighted average would give). In the discussion, the points were raised that in order to reliably increase the precision to the sub-percent level and beyond will likely require not only addressing the scale setting uncertainties (which is reflected in the different values for r<sub>0</sub> obtained by different collaboration and will affect the running of α<sub>s</sub>), but also the inclusion of QED effects.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-9495889932847648752015-09-04T08:40:00.001+01:002015-09-04T08:40:12.053+01:00Fundamental Parameters from Lattice QCD, Day FourToday's first speaker was Andreas Jüttner, who reviewed the extraction of the light-quark CKM matrix elements V<sub>ud</sub> and V<sub>us</sub> from lattice simulations. Since leptonic and semileptonic decay widths of Kaons and pions are very well measured, the matrix element |V<sub>us</sub>| and the ratio |V<sub>us</sub>|/|V<sub>ud</sub>| can be precisely determined if the form factor f<sub>+</sub><sup>Kπ</sup>(0) and the ratio of decay constants f<sub>K</sub>/f<sub>π</sub> are precisely predicted from the lattice. To reach the desired level of precision, the isospin breaking effects from the difference of the up and down quark masses and from electromagnetic interactions will need to be included (they are currently treated in chiral perturbation theory, which may not apply very well in the SU(3) case). Given the required level of precision, full control of all systematics is very important, and the problem of how to properly estimate the associated errors arises, to which different collaborations are offering very different answers. To make the lattice results optimally usable for CKMfitter &Co., one should ideally provide all of the lattice inputs to the CKMfitter fit separately (and not just some combination that presents a particularly small error), as well as their correlations (as far as possible).<br /><br />Unfortunately, I had to miss the second talk of the morning, by Xavier García i Tormo on the extraction of α<sub>s</sub> from the static-quark potential, because our Sonderforschungsbereich (SFB/CRC) is currently up for review for a second funding period, and the local organizers had to be available for questioning by panel members.<br /><br />Later in the afternoon, I returned to the workshop and joined a very interesting discussion on the topic of averaging in the presence of theoretical uncertainties. The large number of possible choices to be made in that context implies that the somewhat subjective nature of systematic error estimates survives into the averages, rather than being dissolved into a consensus of some sort.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-7948577300625744102015-09-04T08:23:00.001+01:002015-09-07T13:11:58.161+01:00Fundamental Parameters from Lattice QCD, Day ThreeToday, our first speaker was Jerôme Charles, who presented new ideas about how treat data with theoretical uncertainties. The best place to read about this is probably his <a href="">talk</a>, but I will try to summarize what I understood. The framework is a firmly frequentist approach to statistics, which answers the basic question of how likely the observed data are if a given null hypothesis is true. In such a context, one can consider a theoretical uncertainty as a fixed bias δ of the estimator under consideration (such as a lattice simulation) which survives the limit of infinite statistics. One can then test the null hypothesis that the true value of the observable in question is μ by constructing a test statistic for the estimator being distributed normally with mean μ+δ and standard deviation σ (the statistical error quoted for the result). The p-value of μ then depends on δ, but not on the quoted systematic error Δ. Since the true value of δ is not known, one has to perform a scan over some region Ω, for example the interval Ω<sub>n</sub>=[-nΔ;nΔ] and take the supremum over this range of δ. One possible extension is to choose Ω adaptively in that a larger range of values needs to be scanned (i.e. a larger true systematic error in comparison to the quoted systematic error is allowed for) for lower p-values; interestingly enough, the resulting curves of p-values are numerically close to what is obtained from a naive Gaussian approach treating the systematic error as a (pseudo-)random variable. For multiple systematic errors, a multidimensional Ω has to be chosen in some way; the most natural choices of a hypercube or a hyperball correspond to adding the errors linearly or in quadrature, respectively. The linear (hypercube) scheme stands out as the only one that guarantees that the systematic error of an average is no smaller than the smallest systematic error of an individual result.<br /><br />The second speaker was Patrick Fritzsch, who gave a nive review of recent lattice determinations of semileptonic heavy-light decays, both the more commonly studied B decays to πℓν and Kℓν, and the decays of the Λ<sub>b</sub> that have recently been investigated by Meinel <i>et al.</i> with the help of LHCb.<br /><br />In the afternoon, both the CKMfitter collaboration and the FLAG group held meetings.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com2tag:blogger.com,1999:blog-8669468.post-88049160794460729332015-09-01T16:29:00.000+01:002015-09-01T16:29:14.736+01:00Fundamental Parameters from Lattice QCD, Day TwoThis morning, we started with a talk by Taku Izubuchi, who reviewed the lattice efforts relating to the hadronic contributions to the anomalous magnetic moment (g-2) of the muon. While the QED and electroweak contributions to (g-2) are known to great precision, most of the theoretical uncertainty presently comes from the hadronic (i.e. QCD) contributions, of which there are two that are relevant at the present level of precision: the contribution from the hadronic vacuum polarization, which can be inserted into the leading-order QED correction, and the contribution from hadronic light-by-light scattering, which can be inserted between the incoming external photon and the muon line. There are a number of established methods for computing the hadronic vacuum polarization, both phenomenologically using a dispersion relation and the experimental R-ratio, and in lattice field theory by computing the correlator of two vector currents (which can, and needs to, be refined in various way in order to achieve competitive levels of precision). No such well-established methods exist yet for the light-by-light scattering, which is so far mostly described using models. There are however, now efforts from a number of different sides to tackle this contribution; Taku mainly presented the appproach by the RBC/UKQCD collaboration, which uses stochastic sampling of the internal photon propagators to explicitly compute the diagrams contributing to (g-2). Another approach would be to calculate the four-point amplitude explicitly (which has recently been done for the first time by the Mainz group) and to decompose this into form factors, which can then be integrated to yield the light-by-light scattering contribution to (g-2).<br /><br />The second talk of the day was given by Petros Dimopoulos, who reviewed lattice determinations of D and B leptonic decays and mixing. For the charm quark, cut-off effects appear to be reasonably well-controlled with present-day lattice spacings and actions, and the most precise lattice results for the D and D<sub>s</sub> decay constants claim sub-percent accuracy. For the b quark, effective field theories or extrapolation methods have to be used, which introduces a source of hard-to-assess theoretical uncertainty, but the results obtained from the different approaches generally agree very well amongst themselves. Interestingly, there does not seem to be any noticeable dependence on the number of dynamical flavours in the heavy-quark flavour observables, as N<sub>f</sub>=2 and N<sub>f</sub>=2+1+1 results agree very well to within the quoted precisions.<br /><br />In the afternoon, the CKMfitter collaboration split off to hold their own meeting, and the lattice participants met for a few one-on-one or small-group discussions of some topics of interest.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-27819583102206747912015-08-31T17:33:00.001+01:002015-08-31T17:34:27.131+01:00Fundamental Parameters from Lattice QCD, Day OneGreetings from Mainz, where I have the pleasure of covering a meeting for you without having to travel from my usual surroundings (I clocked up more miles this year already than can be good from my environmental conscience).<br /><br />Our <a href="http://indico.mitp.uni-mainz.de/conferenceDisplay.py?confId=28">Scientific Programme</a> (which is the bigger of the two formats of meetings that the <a href="http://www.mitp.uni-mainz.de/">Mainz Institute of Theoretical Physics</a> (MITP) hosts, the smaller being Topical Workshops) started off today with two keynote talks summarizing the status and expectations of the <a href="http://itpwiki.unibe.ch/flag/index.php/Review_of_lattice_results_concerning_low_energy_particle_physics">FLAG</a> (Flavour Lattice Averaging Group, presented by Tassos Vladikas) and <a href="http://ckmfitter.in2p3.fr/">CKMfitter</a> (presented by Sébastien Descotes-Genon) collaborations. Both groups are in some way in the business of performing weighted averages of flavour physics quantities, but of course their backgrounds, rationale and methods are quite different in many regards. I will no attempt to give a line-by-line summary of the talks or the afternoon discussion session here, but instead just summarize a few <br />points that caused lively discussions or seemed important in some other way.<br /><br />By now, computational resources have reached the point where we can achieve such statistics that the total error on many lattice determinations of precision quantities is completely dominated by systematics (and indeed different groups would differ at the several-σ level if one were to consider only their statistical errors). This may sound good in a way (because it is what you'd expect in the limit of infinite statistics), but it is also very problematic, because the estimation of systematic errors is in the end really more of an art than a science, having a crucial subjective component at its heart. This means not only that systematic errors quoted by different groups may not be readily comparable, but also that it become important how to treat systematic errors (which may also be correlated, if e.g. two groups use the same one-loop renormalization constants) when averaging different results. How to do this is again subject to subjective choices to some extent. FLAG imposes cuts on quantities relating to the most important sources of systematic error (lattice spacings, pion mass, spatial volume) to select acceptable ensembles, then adds the statistical and systematic errors in quadrature, before performing a weighted average and computing the overall error taking correlations between different results into account using <a href="http://iopscience.iop.org/1402-4896/51/6/002/">Schmelling's procedure</a>. CKMfitter, on the other hand, adds all systematic errors linearly, and uses the <a href="http://arxiv.org/abs/hep-ph/0104062">Rfit procedure</a> to perform a maximum likelihood fit. Either choice is equally permissible, but they are not directly compatible (so CKMfitter can't use FLAG averages as such).<br /><br />Another point raised was that it is important for lattice collaborations computing mixing parameters to not just provide products like <i>f<sub>B</sub>√B<sub>B</sub></i>, but also <i>f<sub>B</sub></i> and <i>B<sub>B</sub></i> separately (as well as information about the correlation between these quantities) in order to help making the global CKM fits easier.<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-17828690812542700762015-07-18T14:19:00.002+01:002015-07-18T14:19:35.642+01:00LATTICE 2015, Day FiveIn a marked deviation from the "standard programme" of the lattice conference series, Saturday started off with parallel sessions, one of which featured my own talk.<br /><br />The lunch break was relatively early, therefore, but first we all assembled in the plenary hall for the conference group photo (a new addition to the traditions of the lattice conference), and was followed by afternoon plenary sessions. The first of these was devoted to finite temperature and density, and started with Harvey Meyer giving the review talk on finite-temperature lattice QCD. The thermodynamic properties of QCD are by now relatively well-known: the transition temperature is agreed to be around 155 MeV, chiral symmetry restoration and the deconfinement transition coincide (as well as that can defined in the case of a crossover), and the number of degrees of freedom is compatible with a plasma of quarks and gluons above the transition, but the thermodynamic potentials approach the Stefan-Boltzmann limit only slowly, indicating that there are strong correlations in the medium. Below the transition, the hadron resonance gas model describes the data well. The Columbia plot describing the nature of the transition as a function of the light and strange quark masses is being further solidified: the size of the lower-left hand corner first-order region is being measured, and the nature of the left-hand border (most likely O(4) second-order) is being explored. Beyond these static properties, real-time properties are beginning to be studied through the finite-temperature spectral functions. One interesting point was that there is a difference between the screening masses (spatial correlation lengths) and quasiparticle masses (from the spectral function) in any given channel, which may even tend in opposite directions as functions of the temperature (as seen for the pion channel).<br /><br />Next, Szabolcs Borsanyi spoke about fluctuations of conserved charges at finite temperature and density. While of course the sum of all outcoming conserved charges in a collision must equal the sum of the ingoing ones, when considering a subvolume of the fireball, this can be best described in the grand canonical ensemble, as charges can move into and out of the subvolume. The quark number susceptibilities are then related to the fluctuating phase of the fermionic determinant. The methods being used to avoid the sign problem include Taylor expansions, fugacity expansions and simulations at imaginary chemical potential, all with their own strengths and weaknesses. Fluctuations can be used as a thermometer to measure the freeze-out temperature.<br /><br />Lastly, Luigi Scorzato reviewed the Lefschetz thimble, which may be a way out of the sign problem (e.g. at finite chemical potential). The Lefschetz thimble is a higher-dimensional generalization of the concept of steepest-descent integration, in which the integral of e<sup>S(z)</sup> for complex S(z) is evaluated by finding the stationary points of S and integrating along the curves passing through them along which the imaginary part of S is constant. On such Lefschetz thimbles, a Langevin algorithm can be defined, allowing for a Monte Carlo evaluation of the path integral in terms of Lefschetz thimbles. In quantum-mechanical toy models, this seems to work already, and there appears hope that this might be a way to avoid the sign problem of finite-density QCD.<br /><br />After the coffee break, the last plenary session turned to physics beyond the Standard Model. Daisuke Kadoh reviewed the progress in putting supersymmetry onto the lattice, which is still a difficult problem due to the fact that the finite differences which replace derivatives on a lattice do not respect the Leibniz rule, introducing SUSY-breaking terms when discretizing. The ways past this are either imposing exact lattice supersymmetries or fine-tuning the theory so as to remove the SUSY-breaking in the continuum limit. Some theories in both two and four dimensions have been simulated successfully, including N=1 Super-Yang-Mills theory in four dimensions. Given that there is no evidence for SUSY in nature, lattice SUSY is of interesting especially for the purpose of verifying the ideas of gauge-dravity duality from the Super-Yang-Mills side, and in one and two dimensions, agreement with the predictions from gauge-gravity duality has been found.<br /><br />The final plenary speaker was Anna Hasenfratz, who reviewed Beyond-the-Standard-Model calculations in technicolor-like theories. If the Higgs is to be a composite particle, there must be some spontaneously broken symmetry that keeps it light, either a flavour symmetry (pions) or a scale symmetry (dilaton). There are in fact a number of models that have a light scalar particle, but the extrapolation of these theories is rendered difficult by the fact that this scalar is (and for phenomenologically interesting models would have to be) lighter than the (techni-)pion, and thus the usual formalism of chiral perturbation theory may not work. Many models of strong BSM interactions have been and are being studied using a large number of different methods, with not always conclusive results. A point raised towards the end of the talk was that for theories with a conformal IR fixed-point, universality might be violated (and there are some indications that e.g. Wilson and staggered fermions seem to give qualitatively different behaviour for the beta function in such cases).<br /><br />The conference ended with some well-deserved applause for the organizing team, who really ran the conference very smoothly even in the face of a typhoon. Next year's lattice conference will take place in Southampton (England/UK) from 24th to 30th July 2016. Lattice 2017 will take place in Granada (Spain).<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-15327223049315393992015-07-17T14:16:00.002+01:002015-07-18T14:19:53.519+01:00LATTICE 2015, Days Three and FourDue to the one-day shift of the entire conference programme relative to other years, Thursday instead of Wednesday was the short day. In the morning, there were parallel sessions. The most remarkable thing to be reported from those (from my point of view) is that MILC are generating a=0.03 fm lattices now, which handily beats the record for the finest lattice spacing; they are observing some problems with the tunnelling of the topological charge at such fine lattices, but appear hopeful that they can be useful.<br /><br />After the lunch break, excursions were offered. I took the trip to Himeji to see Himeji Castle, a very remarkable five-story wooden building that due to its white exterior is also known the "White Heron Castle". During the trip, typhoon Nangka approached, so the rains cut our enjoyment of the castle park a bit short (though seeing koi in a pond with the rain falling into it had a certain special appeal to it, the enjoyment of which I in my Western ignorance suppose might be considered a form of Japanese <i>wabi</i> aesthetics).<br /><br />As the typhoon resolved into a rainstorm, the programme wasn't cancelled or changed, and so today's plenary programme started with a talk on some formal developments in QFT by Mithat Ünsal, who reviewed trans-series, Lefschetz thimbles, and Borel summability as different sides of the same coin. I'm far too ignorant of these more formal field theory topics to do them justice, so I won't try a detailed summary. Essentially, it appears that the expansion of certain theories around the saddle points corresponding to instantons is determined by their expansion around the trivial vacuum, and the ambiguities arising in the Borel resummation of perturbative series when the Borel transform has a pole on the positive real axis can in some way be connected to this phenomenon, which may allow for a way to resolve the ambiguities.<br /><br />Next, Francesco Sannino spoke about the "bright, dark, and safe" sides of the lattice. The bright side referred to the study of visible matter, in particular to the study of technicolor models as a way of implementing the spontaneous breaking of electroweak symmetry, without the need for a fundamental scalar introducing numerous tunable parameters, and with the added benefits of removing the hierarchy problem and the problem of φ<sup>4</sup> triviality. The dark side referred to the study of dark matter in the context of composite dark matter theories, where one should remember that if the visible 5% of the mass of the universe require three gauge groups for their description, the remaining 95% are unlikely to be described by a single dark matter particle and a homogeneous dark energy. The safe side referred to the very current idea of asymptotic safety, which is of interest especially in quantum gravity, but might also apply to some extension of the Standard Model, making it valid at all energy scales.<br /><br />After the coffee break, the traditional experimental talk was given by Toru Iijima of the Belle II collaboration. The Belle II detector is now beginning commissioning at the upcoming SuperKEKB accelerator, which will greatly improved luminosity to allow for precise tests of the Standard Model in the flavour sector. In this, Belle II will be complementary to LHCb, because it will have far lower backgrounds allowing for precision measurements of rare processes, while not being able to access as high energies. Most of the measurements planned at Belle II will require lattice inputs to interpret, so there is a challenge to our community to come up with sufficiently precise and reliable predictions for all required flavour observables. Besides quark flavour physics, Belle II will also search for lepton flavour violation in τ decays, try to improve the phenomenological prediction for (g-2)<sub>μ</sub> by measuring the cross section for e<sup>+</sup>e<sup>-</sup> -> hadrons more precisely, and search for exotic charmonium- and bottomonium-like states.<br /><br />Closely related was the next talk, a review of progress in heavy flavour physics on the lattice given by Carlos Pena. While simulations of relativistic b quarks at the physical mass will become a possibility in the not-too-distant future, for the time being heavy-quark physics is still dominated by the use of effective theories (HQET and NRQCD) and methods based either on appropriate extrapolations from the charm quark mass region, or on the Fermilab formalism, which is sort of in-between. For the leptonic decay constants of heavy-light mesons, there are now results from all formalisms, which generally agree very well with each other, indicating good reliability. For the semileptonic form factors, there has been a lot of development recently, but to obtain precision at the 1% level, good control of all systematics is needed, and this includes the momentum-dependence of the form factors. The z-expansion, and extended versions thereof allowing for simultaneous extrapolation in the pion mass and lattice spacing, has the advantage of allowing for a test of its convergence properties by checking the unitarity bound on its coefficients.<br /><br />After the coffee break, there were parallel sessions again. In the evening, the conference banquet took place. Interestingly, the (excelleent) food was not Japanese, but European (albeit with a slight Japanese twist in seasoning and presentation).<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-31744322578546406912015-07-15T13:47:00.001+01:002015-07-16T03:05:20.505+01:00LATTICE 2015, Day TwoHello again from Lattice 2015 in Kobe. Today's first plenary session began with a review talk on hadronic structure calculations on the lattice given by James Zanotti. James did an excellent job summarizing the manifold activities in this core area of lattice QCD, which is also of crucial phenomenological importance given situations such as the proton radius puzzle. It is now generally agreed that excited-state effects are one of the more important issues facing hadron structure calculations, especially in the nucleon sector, and that these (possibly together with finite-volume effects) are likely responsible for the observed discrepancies between theory and experiment for quantities such as the axial charge of the nucleon. Many groups are studying the charges and form factors of the nucleon, and some have moved on to more complicated quantities, such as transverse momentum distributions. Newer ideas in the field include the use of the Feynman-Hellmann theorem to access quantities that are difficult to access through the traditional three-point-over-two-point ratio method, such as form factors at very high momentum transfer, and quantities with disconnected diagrams (such as nucleon strangeness form factors).<br /><br />Next was a review of progress in light flavour physics by Andreas Jüttner, who likewise gave an excellent overview of this also phenomenologically very important core field. Besides the "standard" quantities, such as the leptonic pion and kaon decay constants and the semileptonic K-to-pi form factors, more difficult light-flavour quantities are now being calculated, including the bag parameter B<sub>K</sub> and other quantities related to both Standard Model and BSM neutral kaon mixing, which require the incorporation of long-distance effects, including those from charm quarks. Given the emergence of lattice ensembles at the physical pion mass, the analysis strategies of groups are beginning to change, with the importance of global ChPT fits receding. Nevertheless, the lattice remains important in determining the low-energy constants of Chiral Perturbation Theory. Some groups are also using newer theoretical developments to study quantities once believed to be outside the purview of lattice QCD, such as final-state photon corrections to meson decays, or the timelike pion form factor.<br /><br />After the coffee break, the Ken Wilson Award for Excellence in Lattice Field Theory was announced. The award goes to Stefan Meinel for his substantial and timely contributions to our understanding of the physics of the bottom quark using lattice QCD. In his acceptance talk, Stefan reviewed his recent work on determining |V<sub>ub</sub>|/|V<sub>cb</sub>| from decays of Λ<sub>b</sub> baryons measured by the LHCb collaboration. There has long been a discrepancy between the inclusive and exclusive (from B -> πlν) determinations of V<sub>ub</sub>, which might conceivably be due to a new (BSM) right-handed coupling. Since LHCb measures the decay widths for Λ<sub>b</sub> to both pμν and Λ<sub>c</sub>μν, combining these with lattice determinations of the corresponding Λ<sub>b</sub> form factors allows for a precise determination of |V<sub>ub</sub>|/|V<sub>cb</sub>|. The results agree well with the exclusive determination from B -> πlν, and fully agree with CKM unitarity. There are, however, still other channels (such as b -> sμ<sup>+</sup>μ<sup>-</sup> and b -> cτν) in which there is still potential for new physics, and LHCb measurements are pending.<br /><br />This was followed by a talk by Maxwell T. Hansen (now a postdoc at Mainz) on three-body observables from lattice QCD. The well-known Lüscher method relates two-body scattering amplitudes to the two-body energy levels in a finite volume. The basic steps in the derivation are to express the full momentum-space propagator in terms of a skeleton expansion involving the two-particle irreducible Bethe-Salpeter kernel, to express the difference between the two-particle reducible loops in finite and infinite volume in terms of two-particle cuts, and to reorganize the skeleton expansion by the number of cuts to reveal that the poles of the propagator (i.e. the energy levels) in finite volume are related to the scattering matrix. For three-particle systems, the skeleton expansion becomes more complicated, since there can now be situations involving two-particle interactions and a spectator particle, and intermediate lines can go on-shell between different two-particle interactions. Treating a number of other technical issues such as cusps, Max and collaborators have been able to derive a Lüscher-like formula three-body scattering in the case of scalar particles with a Z<sub>2</sub> symmetry forbidding 2-to-3 couplings. Various generalizations remain to be explored.<br /><br />The day's plenary programme ended with a talk on the Standard Model prediction for direct CP violation in K-> ππ decays by Christopher Kelly. This has been an enormous effort by the RBC/UKQCD collaboration, who have shown that the ΔI=1/2 rule comes from low-energy QCD by way of strong cancellations between the dominant contributions, and have determined ε' from the lattice for the first time. This required the generation of ensembles with an unusual set of boundary conditions (G-parity boundary conditions on the quarks, requiring complex conjugation boundary conditions on the gauge fields) in space to enforce a moving pion ground state, as well as the precise evaluation of difficult disconnected diagrams using low modes and stochastic estimators, and treatment of finite-volume effects in the Lellouch-Lüscher formalism. Putting all of this together with the non-perturbative renormalization (in the RI-sMOM scheme) of ten operators in the electroweak Hamiltonian gives a result which currently still has three times the experimental error, but is systematically improvable, with better-than-experimental precision expected in maybe five years.<br /><br />In the afternoon there were parallel sessions again, and in the evening, the poster session took place. Food ran out early, but it was pleasant to see <a href="http://arxiv.org/abs/1306.1440">free-form smearing</a> begin improved upon and used to very good effect by Randy Lewis, Richard Woloshyn and students.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-19455138267149688142015-07-14T12:30:00.000+01:002015-07-14T12:30:28.925+01:00LATTICE 2015, Day OneHello from Kobe, where I am attending the Lattice 2015 conference. The trip here was uneventful, as was the jetlag-day.<br /><br />The conference started yesterday evening with a reception in the Kobe Animal Kingdom (there were no animals when we were there, though, with the exception of some fish in a pond and some cats in a cage, but there were lot of plants).<br /><br />Today, the scientific programme began with the first plenary session. After a welcome address by Akira Ukawa, who reminded us of the previous lattice meetings held in Japan and the tremendous progress the field has made in the intervening twelve years, Leonardo Giusti gave the first plenary talk, speaking about recent progress on chiral symmetry breaking. Lattice results have confirmed the proportionality of the square of the pion mass to the quark mass (i.e. the Gell-Mann-Oakes-Renner (GMOR) relation, a hallmark of chiral symmetry breaking) very accurately for a long time. Another relation involving the chiral condensate is the Banks-Casher relation, which relates it to the eigenvalue density of the Dirac operator at zero. It can be shown that the eigenvalue density is renormalizable, and that thus the mode number in a given interval is renormalization-group invariant. Two recent lattice studies, one with twisted-mass fermions and one with O(a)-improved Wilson fermions, confirm the Banks-Casher relation, with the chiral condensates found agreeing very well with those inferred from GMOR. Another relation is the Witten-Veneziano relation, which relates the η' mass to the topological susceptibility, thus explaining how precisely the η' is not a Goldstone boson. The topological charge on the lattice can be defined through the index of the Neuberger operator or through chain of spectral porjectors, but a recently invented and much cheaper definition is through the topological charge density at finite flow time in Lüscher's Wilson flow formalism. The renormalization properties of the Wilson flow allow for a derivation of the universality of the topological susceptibility, and numerical tests using all three definitions indeed agree within errors in the continuum limit. Higher cumulants determined in the Wilson flow formalism agree with large-N<sub>c</sub> predictions in pure Yang-Mills, and the suppression of the topological susceptibility in QCD relative to the pure Yang-Mills case is in line with expectations (which in principle can be considered an <i>a posteriori</i> determination of N<sub>f</sub> in agreement with the value used in simulations).<br /><br />The next speaker was Yu Nakayama, who talked about a related topic, namely the determination of the chiral phase transition in QCD from the conformal bootstrap. The chiral phase transition can be studied in the framework of a Landau effective theory in three dimensions. While the mean-field theory predicts a second-order phase transition in the O(4) universality class, one-loop perturbation theory in 4-ε dimensions predicts a first-order phase transition at ε=1. Making use of the conformal symmetry of the effective theory, one can apply the conformal bootstrap method, which combines an OPE with crossing relations to obtain results for critical exponents, and the results from this method suggest that the phase transition is in fact of second order. This also agrees with many lattice studies, but others disagree. The role of the anomalously broken U(1)<sub>A</sub> symmetry in this analysis appears to be unclear.<br /><br />After the coffee break, Tatsumi Aoyama, a long-time collaborator in the heroic efforts of Kinoshita to calculate the four- and five-loop QED contributions to the electron and muon anomalous moments, gave a plenary talk on the determination of the QED contribution to lepton (g-2). For likely readers of this blog, the importance of (g-2) is unlikely to require an explanation: the current 3σ tension between theory and experiment for (g-2)<sub>μ</sub> is the strongest hint of physics beyond the Standard Model so far, and since the largest uncertainties on the theory side are hadronic, lattice QCD is challenged to either resolve the tension or improve the accuracy of the predictions to the point where the tension becomes an unambiguous, albeit indirect, discovery of new physics. The QED calculations are on the face of it simpler, being straightforward Feynman diagram evaluations. However, the number of Feynman diagrams grows so quickly at higher orders that automated methods are required. In fact, in a first step, the number of Feynman diagrams is reduced by using the Ward-Takahashi identity to relate the vertex diagrams relevant to (g-2) to self-energy diagrams, which are then subjected to an automated renormalization procedure using the Zimmermann forest formula. In a similar way, infrared divergences are subtracted using a more complicated "annotated forest"-formula (there are two kinds of IR subtractions needed, so the subdiagrams in a forest need to be labelled with the kind of subtraction). The resulting UV- and IR-finite integrands are then integrated using VEGAS in Feynman parameter space. In order to maintain the required precision, quadruple-precision floating-point numbers (or an emulation thereof) must be used. Whether these methods could cope with the six-loop QED contribution is not clear, but with the current and projected experimental errors, that contribution will not be required for the foreseeable future, anyway.<br /><br />This was followed by another (g-2)-related plenary, with Taku Izubichi speaking about the determination of anomalous magnetic moments and nucleon electric dipole moments in QCD. In particular the anomalous magnetic moment has become such an active topic recently that the time barely sufficed to review all of the activity in this field, which ranges from different approaches to parameterizing the momentum dependence of the hadronic vacuum polarization, through clever schemes to reduce the noise by subtracting zero-momentum contributions, to new ways of extracting the vacuum polarization through the use of background magnetic fields, as well as simulations of QCD+QED on the lattice. Among the most important problems are finite-volume effects.<br /><br />After the lunch break, there were parallel sessions in the afternoon. I got to chair the first session on hadron structure, which was devoted to determinations of hadronic contributions to (g-2)<sub>μ</sub>.<br /><br />After the coffee break, there were more parallel sessions, another complete one of which was devoted to (g-2) and closely-related topics. A talk deserving to be highlighted was given by Jeremy Green, who spoke about the first direct calculation of the hadronic light-to-light scattering amplitude from lattice QCD.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-36177784996893429362015-04-10T09:19:00.000+01:002015-04-10T15:59:59.883+01:00Workshop "Fundamental Parameters from Lattice QCD" at MITP (upcoming deadline)Recent years have seen a significant increase in the overall accuracy of lattice QCD calculations of various hadronic observables. Results for quark and hadron masses, decay constants, form factors, the strong coupling constant and many other quantities are becoming increasingly important for testing the validity of the Standard Model. Prominent examples include calculations of Standard Model parameters, such as quark masses and the strong coupling constant, as well as the determination of CKM matrix elements, which is based on a variety of input quantities from experiment and theory. In order to make lattice QCD calculations more accessible to the entire particle physics community, several initiatives and working groups have sprung up, which collect the available lattice results and produce global averages.<br /><br />The scientific programme "<a href="https://indico.mitp.uni-mainz.de/conferenceDisplay.py?confId=28">Fundamental Parameters from Lattice QCD</a>" at the Mainz Institute of Theoretical Physics (<a href="http://www.mitp.uni-mainz.de/">MITP</a>) is designed to bring together lattice practitioners with members of the phenomenological and experimental communities who are using lattice estimates as input for phenomenological studies. In addition to sharing the expertise among several communities, the aim of the programme is to identify key quantities which allow for tests of the CKM paradigm with greater accuracy and to discuss the procedures in order to arrive at more reliable global estimates.<br /><br />The deadline for <a href="https://indico.mitp.uni-mainz.de/confRegistrationFormDisplay.py/display?confId=28" title="Registration form">registration</a> is <b>Wednesday, 15 April 2015</b>. Please register <a href="https://indico.mitp.uni-mainz.de/confRegistrationFormDisplay.py/display?confId=28" title="Register now!">at this link</a>.Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-62910768935813013992015-03-12T17:01:00.004+00:002015-03-12T17:07:29.688+00:00QNP 2015, Day Five<i>Apologies for the delay in posting this. Travel and jetlag kept me from attending to it earlier.</i><br /><br />The first talk today was by Guy de Teramond, who described applications of light-front superconformal quantum mechanics to hadronic physics. I have to admit that I couldn't fully take in all the details, but as far as I understood an isomorphy between AdS<sup>2</sup> and the conformal group in one dimension can be used to derive a form of the light-front Hamiltonian for mesons from an AdS/QCD correspondence, in which the dilaton field is fixed to be φ(z)=1/2 z<sup>2</sup> by the requirement of conformal invariance, and a similar construction in the superconformal case leads to a light-front Hamiltonian for baryons. A relationship between the Regge trajectories for mesons and baryons can then be interpreted as a form of supersymmetry in this framework.<br /><br />Next was Beatriz Gay Ducati with a review of the pheonomenology of heavy quarks in nuclear matter, a topic where there are still many open issues. The photoproduction of quarkonia on nucleons and nuclei allows to probe the gluon distribution, since the dominant production process is photon-gluon fusion, but to be able to interpret the data, many nuclear matter effects need to be understood.<br /><br />After the coffee break, this was followed by a talk by Hrayr Matevosyan on transverse momentum distributions (TMDs), which are complementary to GPDs in the sense of being obtained by integrating out other variables starting from the full Wigner distributions. Here, again, there are many open issues, such as the Sivers, Collins or Boer-Mulders effects.<br /><br />The next speaker was Raju Venugopalan, who spoke about two outstanding problems in QCD at high parton densities, namely the question of how the systems created in heavy-ion collisions thermalise, and the phenomenon of "the ridge" in proton-nucleus collisions, which would seem to suggest hydrodynamic behaviour in a system that is too small to be understood as a liquid. Both problems may have to do with the structure of the dense initial state, which is theorised to be a colour-glass condensate or "glasma", and the way in which it evolves into a more dilute system.<br /><br />After the lunch break, Sonny Mantry reviewed some recent advances made in applying Soft-Collinear Effective Theory (SCET) to a range of questions in strong-interaction physics. SCET is the effective field theory obtained when QCD fluctuations around a hard particle momentum are considered to be small and a corresponding expansion (analogous to the 1/m expansion in HQET) is made. SCET has been successfully applied to many different problems; an interesting and important one is the problem of relating the "Monte Carlo mass" usually quoted for the top quark to the top quark mass in a more well-defined scheme such as MSbar.<br /><br />The last talk in the plenary programme was a review of the Electron-Ion Collider (EIC) project by Zein-Eddine Meziani. By combining the precision obtainable using an electron beam with the access to the gluon-dominated regime provided by a havy ion beam, as well as the ability to study the nucleon spin using a polarised nucleon beam, the EIC will enable a much more in-depth study of many of the still unresolved questions in QCD, such as the nucleon spin structure and colour distributions. There are currently two competing designs, the eRHIC at Brookhaven, and the MEIC at Jefferson Lab.<br /><br />Before the conference closed, Michel Garçon announced that the next conference of the series (QNP 2018) will be held in Japan (either in Tsukuba or in Mito, Ibaraki prefecture). The local organising committee and conference office staff received some well-deserved applause for a very smoothly-run conference, and the scientific part of the conference programme was adjourned.<br /><br />As it was still in the afternoon, I went with some colleagues to visit <a href="http://es.wikipedia.org/wiki/La_Sebastiana">La Sebastiana</a>, the house of <a href="http://en.wikipedia.org/wiki/Pablo_Neruda">Pablo Neruda</a> in Valparaíso, taking one of the city's famous <i>ascensores</i> down (although up might have been more convenient, as the streets get very steep) before walking back to Viña del Mar along the sea coast.<br /><br />The next day, there was an organised excursion to a vineyard in the Casablanca valley, where we got to taste some very good Chilean wines (some of the them matured in traditional clay vats) and liqueurs with a very pleasant lunch.<br /><br />I got to spend another day in Valparaíso before travelling back (a happily uneventful, if again rather long trip).<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-4624295650904595602015-03-06T12:13:00.003+00:002015-03-06T12:15:50.731+00:00QNP 2015, Day FourThe first talk today was a review of experimental results in light-baryon spectroscopy by Volker Credé. While much progress has been made in this field, in particular in the design of so-called complete experiments, which as far as I understand measure multiple observables to unambiguously extract a complete description of the amplitudes for a certain process, there still seem to be surprisingly many unknowns. In particular, the fits to pion photoproduction in doubly-polarised processes seem to disagree strongly between different descriptions (such as MAID).<br /><br />Next was Derek Leinweber with a review of light hadron spectroscopy from the lattice. The <i>de facto</i> standard method in this field is the variational method (GEVP), although there are some notable differences in how precisely different groups apply it (e.g. solving the GEVP at many times and fitting the eigenvalues vs. forming projected correlators with the eigenvectors of the GEVP solved at a single time -- there are proofs of good properties for the former that don't exist for the latter). The way in which the basis of operators for the GEVP is build is also quite different as used by different groups, ranging from simply using different levels of quark field smearing to intricate group-theoretic constructions of multi-site operators. There are also attempts to determine how much information can be extracted from a given set of correlators, e.g. recently by the <a href="http://arxiv.org/abs/1411.6765">Cyprus/Athens group</a> using Monte Carlo simulations to probe the space of fitting parameters (a loosely related older idea based on <a href="http://arxiv.org/abs/0707.2788">evolutionary fits</a> wasn't mentioned).<br /><br />This was followed by a talk by Susan Gardner about testing fundamental symmetries with quarks. While we know that there must be physics beyond the Standard Model (because the SM does not explain dark matter, nor does it provide enough CP violation to explain the observed baryon asymmetry), there is so far no direct evidence of any BSM particle. Low-energy tests of the SM fall into two broad categories: null tests (where the SM predicts an exact null result, as for violations of B-L) and precision tests (where the SM prediction can be calculated to very high accuracy, as for (g-2)<sub>μ</sub>). Null tests play an important role in so far as they can be used to impose a lower limit for the BSM mass scale, but many of them are atomic or nuclear tests, which have complicated theory errors. The currently largest tensions indicating a possible failure of the Standard Model to describe all observations are the proton radius puzzle, and (g-2)<sub>μ</sub>. A possible explanation of either or both of those in terms of a "dark photon" is on the verge of being ruled out, however, since most of the relevant part of the mass/coupling plane has already been excluded by dark photon searches, and the rest of it will soon be (or else the dark photon will be discovered). Other tests in the hadronic sector, which seem to be less advanced so far, are the search for non-(V-A) terms in β-decays, and the search for neutron-antineutron oscillations.<br /><br />After the coffee break and the official conference photo, Isaac Vidaña took the audience on a "half-hour walk through the physics of neutron stars". Neutron stars are both almost-black holes (whose gravitation must be described in General Relativity) and extremely massive nuclei (whose internal dynamics must be described using QCD). Observations of binary pulsars allow to determine the masses of neutron stars, which are found to range up to at least two solar masses. However, the Tolman-Oppenheimer-Volkov equations for the stability of neutron stars lead to a maximum mass for a neutron star that depends on the equation of state of the nuclear medium. The observed masses severely constrain the equation of state and in particular seem to exclude models in which hyperons play an important role; however, it seems to be generally agreed that hyperons must play an important role in neutron stars, leading to a "hyperon puzzle", the solution of which will require an improved understanding of the structure and interactions of hyperons.<br /><br />The last plenary speaker of the day was Stanley Brodsky with the newest developments from light-front holography. The light-front approach, which has in the past been very successful in (1+1)-dimensional QCD, is based on the front form of the Hamiltonian formalism, in which a light-like, rather than a timelike, direction is chosen as the normal defining the Cauchy surfaces on which initial data are specified. In the light-front Hamiltonian approach, the vacuum of QCD is trivial and the Hilbert space can be constructed as a straightforward Fock space. With some additional ansätze taken from AdS/CFT ideas, QCD is reduced to a Schrödinger-like equation for the light-cone wavefunctions, from which observables are extracted. Apparently, all known observations are described perfectly in this approach, but (as for the Dyson-Schwinger or straight AdS/QCD approaches) I do not understand how systematic errors are supposed to be quantified.<br /><br />In the afternoon there were parallel talks. An interesting contribution was given by Mainz PhD student Franziska Hagelstein, who demonstrated how even a very small non-monotonicity in the electric form factor at low Q<sup>2</sup> (where there are no ep scattering data) could explain the difference between the muonic and electronic hydrogen results for the proton radius.<br /><br />The conference banquet took place in the evening at a very nice restaurant, and fun was had over cocktails and an excellent dinner.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-1141064874439987972015-03-05T12:27:00.005+00:002015-03-05T12:27:54.534+00:00QNP 2015, Day ThreeToday began with a talk by Mikhail Voloshin on QCD sum rules and heavy-quark states. The idea of exploiting quark-hadron duality to link perturbatively calculable current-current correlators to hadronic obervables and extract mesonic decay constants or quark masses is quite old, but has received a boost in recent years with the advent of three- and four-loop perturbative calculations in particularly from Chetyrkin and collaborators, which have also been used in conjunction with lattice results, e.g. by the HPQCD collaboration.<br /><br />A review of hadron spectroscopy at B factories (including LHCb) by Roberto Mussa followed. The charmonium and bottomonium spectra are now measured to great detail, with recent additions being 1D and 3P states, and more states are also being discovered in the heavy-light (where the B<sub>c</sub>(2S) has recently been discovered at ATLAS) and heavy-quark baryon (where the most recent discovery was the Ξ<sub>b</sub>) sectors, and many more transitions being discovered and studied.<br /><br />The next speaker was Raphaël Dupré, who spoke about colour propagation and neutralisation in strongly interacting systems. The idea here appears to be that in hadronisation processes, quarks first loose energy by radiating gluons and thus turn into colourless pre-hadrons, which then bind into hadrons on a longer timescale, and there seems to be experimental evidence supporting this energy-loss model.<br /><br />After the coffee break, Javier Castillo reviewed quarkonium suppression and regeneration in heavy-ion collisions. Quarkonia are generally considered important probes of the quark-gluon plasma, because the production of heavy quark-antiquark pairs is a perturbative process that happens at high energies early in the collision, while their binding is non-perturbative and is expected to be suppressed by Debye screening in the coloured plasma. As a consequence, more tightly bound quarkonia, like the Y(1S), can exist at higher temperatures, while the more lightly bound charmonia or Y(3S) states will "melt" at lower temperatures. However, quarkonia can also be regenerated by thermalised heavy quarks rejoining into quarkonia at the phase boundary. Experimental data support the screening picture, with the J/ψ being more suppressed at the LHC than at STAR (because of the higher temperature), the Y(2S) more suppressed than the Y(1S), and transport models with a negligible regeneration component describing the data well. The regeneration component increases at low p<sub>T</sub>, and the elliptic flow of the charm quarks is inherited by the regenerated J/ψ mesons. Some more difficult to understand effects of the nuclear environment, called Cold Nuclear Matter (CNM) effects are beginning to be seen in the data.<br /><br />Next was Zoltan Fodor with a talk about Lattice QCD results at zero and finite temperature from the BMW collaboration. By simulating QCD+QED with 1+1+1+1 flavours of dynamical quarks, BMW have been able to determine the isospin splitting of the nucleon and other baryonic systems. This work, which appears set to become a cover story in "Science", had to overcome a number of serious obstacles, in particular long-range autocorrelations (which could cured by a Fourier-accelerated HMC variant) and power-law finite-volume effects (which had to be fitted to results obtained at a range of volumes) introduced by the massless photon. In the finite-temperature regime, the crossover temperature is now generally agreed to be around 150-160 MeV, but the position and even existence of the critical endpoint is still contentious (and any existing results are not yet continuum-extrapolated in any case).<br /><br />After the lunch break, Yiota Foka gave an overview of heavy-ion results from RHIC and the LHC. The elliptic flow is still found to be in agreement with perfect hydrodynamics, but people are now also studying higher harmonics, as well as the interplay between jets and flow, which provide important constraints on the physics of the quark-gluon plasma. At the LHC, it has been found that it is the mass, and not the valence quark content, that drives the flow behaviour of hadrons, as the φ meson has the same flow behaviour as the proton.<br /><br />The next speaker was Carl Gagliardi, who reviewed results in nucleon structure from high-energy polarised proton-proton collisions. Proton-proton scattering is complementary to DIS in that it gives access to the gluonic degrees of freedom which are invisible to electrons, and RHIC has a programme of polarised proton collisions to explore the spin structure of the nucleon. Without the RHIC data, the gluon polarisation ΔG is almost unconstrained, but with the RHIC data, it is seen to be clearly positive and contribute about 0.2 to the proton spin. Using W production, it is possible to separate polarised quark and antiquark distributions, and there is more to come in the near future.<br /><br />The last plenary speaker of the day was Craig Roberts, who reviewed the pion and nucleon structure from the point of view of the Dyson-Schwinger equations approach. In this approach, the pion is closely linked to the quark mass function, which comes out of a quark gap equation and describes how the running quark mass at high energies turns into a much larger constituent quark mass at low energies. Landau-gauge gluons also become massive at low energies, and confinement is explained as the splitting of poles into pairs of conjugate complex poles giving an exponentially damped behaviour of the position space propagator. While this approach seems to be able to readily explain every single known experimental result, I do not understand how the systematic errors from the truncation of the infinite tower of DSEs are supposed to be controlled or quantified.<br /><br />After the coffee break, there were parallel sessions. An interesting parallel talk was given by Johan Bijnens, who has determined the leading logarithms for the nucleon mass (and some other systems) to rather high orders (which also for effective theories can be done using only one-loop integrals from a consistency argument by Weinberg).<br /><br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-30841792030095928982015-03-04T12:05:00.003+00:002015-03-04T12:06:26.803+00:00QNP 2015, Day TwoHello again from Valparaíso. Today's first speaker was Johan Bijnens with a review of recent results from chiral perturbation theory in the mesonic sector, including recent results for charged pion polarisabilities and for finite-volume corrections to lattice measurements. To allow others to perform their own calculations for their own specific needs (which might include technicolor-like theories, which will generally have different patterns of chiral symmetry breaking, but otherwise work just the same way), Bijnens & Co. have recently published CHIRON, a general two-loop mesonic χPT package. The leading logarithms have been determined to high orders, and it has been found that the speed of convergence depends both on the observable and on whether the leading-order or physical pion decay constant is used.<br /><br />Next was Boris Grube, who presented some recent results from light-meson spectroscopy. The light mesons are generally expected to be some kind of superpositions of quark-model states, hybrids, glueballs, tetraquark and molecular states, as may be compatible with their quantum numbers in each case. The most complex sector is the 0<sup>++</sup> sector of f<sub>0</sub> mesons, in which the lightest glueball state should lie. While the γγ width of the f<sub>0</sub>(1500) appears to be compatible with zero, which would agree with the expectations for a glueball, whereas the f<sub>0</sub>(1710) has a photonic width more in agreement with being an s-sbar state, in J/ψ -> γ (ηη), which as a gluon-rich process should couple strongly to glueball resonances, little or no f<sub>0</sub>(1500) is seen, whereas a glueball nature for the f<sub>0</sub>(1710) would be supported by these results. New data to come from GlueX, and later from PANDA, should help to clarify things.<br /><br />The next speaker was Paul Sorensen with a talk on the search for the critical point in the QCD phase diagram. The quark-gluon plasma at RHIC is not only a man-made system that is over 300 times hotter than the centre of the Sun, it is also the most perfect fluid known, as it close to saturates the viscosity bound η/s > 1/(4π). Studying it experimentally is quite difficult, however, since one must extrapolate back to a small initial fireball, or "little bang", from correlations between thousands of particle tracks in a detector, not entirely dissimilar from the situation in cosmology, where the properties of the hot big bang (and previous stages) are inferred from angular correlations in the cosmic microwave background. Beam energy scans find indications that the phase transition becomes first-order at higher densities, which would indicate the existence of a critical endpoint, but more statistics and more intermediate energies are needed.<br /><br />After the coffee break, François-Xavier Girod spoke about Generalised Parton Distributions (GPDs) and deep exclusive processes. GPDs, which reduce to form factors and to parton distributions upon integrating out the unneeded variables in each case, correspond to a three-dimensional image of the nucleon performed in the longitudinal momentum fraction and the transverse impact parameter, and their moments are related to matrix elements of the energy-momentum tensor. Experimentally, they are probed using deeply virtual Compton scattering (DVCS); the 12 GeV upgrade at Jefferson Lab will increase the coverage in both Bjørken-x and Q<sup>2</sup>, and the planned electron-ion collider is expected to allow probing the sea and gluon GPDs as well.<br /><br />After the lunch break, there were parallel sessions. I chaired the parallel session on lattice and other perturbative methods, with presentations of lattice results by Eigo Shintani and Tereza Mendes, as well as a number of AdS/QCD-related results by various others.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-52685347929355549202015-03-03T14:38:00.001+00:002015-03-03T14:38:23.203+00:00QNP 2015, Day OneHello from Valparaíso, where I continue this year's hectic conference circuit at the 7th International Conference on Quarks and Nuclear Physics (QNP 2015). Except for some minor inconveniences and misunderstandings, the long trip to Valparaíso (via Madrid and Santiago de Chile) went quite smoothly, and so far, I have found Chile a country of bright sunlight and extraordinarily helpful and friendly people.<br /><br />The first speaker of the conference was Emanuele Nocera, who reviewed nucleon and nuclear parton distributions. The study of parton distributions become necessary because hadrons are really composed not simply of valence quarks, as the quark model would have it, but of an indefinite number of (sea) quarks, antiquarks and gluons, any of which can contribute to the overall momentum and spin of the hadron. In an operator product expansion framework, hadronic scattering amplitudes can then be factorised into Wilson coefficients containing short-distance (perturbative) physics and parton distribution functions containing long-distance (non-perturbative) physics. The evolution of the parton distribution functions (PDFs) with the momentum scale is given by the DGLAP equations containing the perturbatively accessible splitting functions. The PDFs are subject to a number of theoretical constraints, of which the sum rules for the total hadronic momentum and valence quark content are the most prominent. For nuclei, on can assume that a similar factorisation as for hadrons still holds, and that the nuclear PDFs are linear combinations of nucleon PDFs modified by multiplication with a binding factor; however, nuclei exhibit correlations between nucleons, which are not well-described in such an approach. Combining all available data from different sources, global fits to PDFs can be performed using either a standard χ<sup>2</sup> fit with a suitable model, or a neural network description. There are far more and better data on nucleon than nuclear PDFs, and for nucleons the amount and quality of the data also differs between unpolarised and polarised PDFs, which are needed to elucidate the "proton spin puzzle".<br /><br />Next was the first lattice talk of the meeting, given by Huey-Wen Lin, who gave a review of the progress in lattice studies of nucleon structure. I think Huey-Wen gave a very nice example by comparing the computational and algorithmic progress with that in videogames (I'm not an expert there, but I think the examples shown were screenshots of Nethack versus some modern first-person shooter), and went on to explain the importance of controlling all systematic errors, in particular excited-state effects, before reviewing recent results on the tensor, scalar and axial charges and the electromagnetic form factors of the nucleon. As an outlook towards the current frontier, she presented the inclusion of disconnected diagrams and a new idea of obtaining PDFs from the lattice more directly rather than through their moments.<br /><br />The next speaker was Robert D. McKeown with a review of JLab's Nuclear Science Programme. The CEBAF accelerator has been upgraded to 12 GeV, and a number of experiments (GlueX to search for gluonic excitations, MOLLER to study parity violation in Møller scattering, and SoLID to study SIDIS and PVDIS) are ready to be launched. A number of the planned experiments will be active in areas that I know are also under investigation by experimental colleagues in Mainz, such as a search for the "dark photon" and a study of the running of the Weinberg angle. Longer-term plans at JLab include the design of an electron-ion collider.<br /><br />After a rather nice lunch, Tomofumi Nagae spoke about the hadron physics programme an J-PARC. In spite of major setbacks by the big earthquake and a later radiation accident, progress is being made. A search for the Θ<sup>+</sup> pentaquark did not find a signal (which I personally do not find surprising, since the whole pentaquark episode is probably of more immediate long-term interest to historians and sociologists of science than to particle physicists), but could not completely exclude all of the discovery claims.<br /><br />This was followed by a take by Jonathan Miller of the MINERνA collaboration presenting their programme of probing nuclei with neutrinos. Major complications include the limited knowledge of the incoming neutrino flux and the fact that final-state interactions on the nuclear side may lead to one process mimicking another one, making the modelling in event generators a key ingredient of understanding the data.<br /><br />Next was a talk about short-range correlations in nuclei by Or Henn. Nucleons subject to short-range correlations must have high relative momenta, but a low center-of-mass momentum. The experimental studies are based on kicking a proton out of a nucleus with an electron, such that both the momentum transfer (from the incoming and outgoing electron) and the final momentum of the proton are known, and looking for a nucleon with a momentum close to minus the difference between those two (which must be the initial momentum of the knocked-out proton) coming out. The astonishing result is that at high momenta, neutron-proton pairs dominate (meaning that protons, being the minority, have a much larger chance of having high momenta) and are linked by a tensor force. Similar results are known from other two-component Fermi systems, such as ultracold atomic gases (which are of course many, many orders of magnitude less dense than nuclei).<br /><br />After the coffee break, Heinz Clement spoke about dibaryons, specifically about the recently discovered d<sup>*</sup>(2380) resonance, which taking all experimental results into account may be interpreted as a ΔΔ bound state<br /><br />The last talk of the day was by André Walker-Loud, who reviewed the study of nucleon-nucleon interactions and nuclear structure on the lattice, starting with a very nice review of the motivations behind such studies, namely the facts that big-bang nucleosynthesis is very strongly dependent on the deuterium binding energy and the proton-neutron mass difference, and this fine-tuning problem needs to be understood from first principles. Besides, currently the best chance for discovering BSM physics seems once more to lie with low-energy high-precision experiments, and dark matter searches require good knowledge of nuclear structure to control their systematics. Scattering phase shifts are being studied through the Lüscher formula. Current state-of-the-art studies of bound multi-hadron systems are related to dibaryons, in particular the question of the existence of the H-dibaryon at the physical pion mass (note that the dineutron, certainly unbound in the real world, becomes bound at heavy enough pion masses), and three- and four-nucleon systems are beginning to become treatable, although the signal-to-noise problem gets worse as more baryons are added to a correlation function, and the number of contractions grows rapidly. Going beyond masses and binding energies, the new California Lattice Collaboration (CalLat) has preliminary results for hadronic parity violation in the two-nucleon system, albeit at a pion mass of 800 MeV.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-35069385127172395652015-02-27T10:01:00.000+00:002015-02-27T10:01:06.061+00:00Back from MumbaiOn Saturday, my last day in Mumbai, a group of colleagues rented a car with a driver to take a trip to Sanjay Gandhi National Park and visit the <a href="http://en.wikipedia.org/wiki/Kanheri_caves">Kanheri caves</a>, a Buddhist site consisting of a large number of rather simple monastic cells and some worship and assembly halls with ornate reliefs and inscriptions, all carved out out of solid rock (some of the cell entrances seem to have been restored using steel-reinforced concrete, though).<br /><br />On the way back, we stopped at <a href="http://en.wikipedia.org/wiki/Mani_Bhavan">Mani Bhavan</a>, where Mahatma Gandhi lived from 1917 to 1934, and which is now a museum dedicated to his live and legacy.<br /><br />In the night, I flew back to Frankfurt, where the temperature was much lower than in Mumbai; in fact, on Monday there was snow.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0tag:blogger.com,1999:blog-8669468.post-5098186261802833472015-02-20T12:02:00.002+00:002015-02-20T12:02:43.470+00:00Perspectives and Challenges in Lattice Gauge Theory, Day FiveToday's programme started with a talk by Santanu Mondal on baryons in the sextet gauge model, which is a technicolor-style SU(3) gauge theory with a doublet of technifermions in the sextet (two index symmetric) representation, and a minimal candidate for a technicolor-like model with an IR almost-fixed point. Using staggered fermions, he found that when setting the scale by putting the technipion's decay constant to the value derived from identifying the Higgs vacuum expectation value as the technicondensate, the baryons had masses in excess of 3 TeV, heavy enough to not yet have been discovered by the LHC, but to be within reach of the next run. However, the anomaly cancellation condition when embedding the theory into the Standard Model of the electroweak interactions requires charge assignments such that the lightest technibaryon (which would be a stable particle) would have a fractional electrical charge of 1/2, and while the cosmological relic density can be made small enough to evade detection, the technibaryons produced by the cosmic rays in the Earth's atmosphere should have been able to accumulate (there currently appear to be no specific experimental exclusions for charge-1/2 particles though).<br /><br />Next was Nilmani Mathur speaking about mixed action simulations using overlap valence quarks on the MILC HISQ ensembles (which include the radiative <a href="http://arxiv.org/abs/0812.0503">corrections</a> to the lattice gluon action from the quarks). Tuning the charm quark mass via the kinetic rather than rest mass of charmonium, the right charmonium hyperfine splitting is found, as well as generally correct charmonium spectra. Heavy-quark baryons (up to and including the Ω<sub>ccc</sub>) have also been simulated, with results in good agreement with experimental ones where the latter exist. The mixed-action effects appear to be mild small in mixed-action χPT, and only half as large as those for domain-wall valence fermions on an asqtad sea.<br /><br />In a brief note, Gunnar Bali encouraged the participants of the workshop to seek out opportunities for Indo-German research collaboration, of which there are still only a limited number of instances.<br /><br />After the tea break, there were two more theoretical talks, both of them set in the framework of Hamiltonian lattice gauge theory: Indrakshi Raychowdhury presented a loop formulation of SU(2) lattice gauge theory based on the prepotential formalism, where both the gauge links and their conjugate electrical fields are constructed from harmonic oscillator variables living on the sites using the Schwinger construction. By some ingenious rearrangements in terms of "fusion variables", a representation of the perturbative series for Hamiltonian lattice gauge theory purely in terms of integer-valued quantum numbers in a geometric-combinatorial construction was derived.<br /><br />Lastly, Sreeraj T.P. presented a derivation of an analogy between the Gauss constraint in Hamiltonian lattice gauge theory and the condition of equal "angular impulses" in the SU(2) x SU(2) description of the SO(4) symmetry of the Coulomb potential to derive a description of the Hilbert space of SU(2) lattice gauge theory in terms of hydrogen atom (n,l,m) variables located on the plaquettes subject only to the global constraint of vanishing total angular momentum, from where a variational ansatz for the ground state can be constructed.<br /><br />The workshop closed with some well-deserved applause for the organisers and all of the supporting technical and administrative staff, who have ensured that this workshop ran very smoothly indeed. Another excellent lunch (I understand that our lunches have been a kind of culinary journey through India, starting out in the north on Monday and ending in Kerala today) concluded the very interesting workshop.<br /><br />I will keep the small subset of my readers whom it may interest updated about my impressions from an excursion planned for tomorrow and my trip back.<br />Georg v. Hippelhttp://www.blogger.com/profile/15355104409462033075noreply@blogger.com0