It is often said that “nobody understands quantum mechanics” – a phrase usually attributed to Richard Feynman. This statement may, however, be misleading to the uninitiated. There is certainly a high level of understanding of quantum mechanics. The point, moreover, is that there is more than one way to understand the theory, and each of these ways requires us to make some disturbing concessions.
Carlo Rovelli’s Helgoland is therefore a welcome popular book – a well-written and easy-to-follow exploration of quantum mechanics and its interpretation. Rovelli is a theorist working mainly on quantum gravity and foundational aspects of physics. He is also a very successful popular author, distinguished by his erudition and his ability to illuminate the bigger picture. His latest book is no exception.
Helgoland is a barren German island of the North Sea where Heisenberg co-invented quantum mechanics in 1925 while on vacation. The extraordinary sequence of events between 1925 and 1926, when Heisenberg, Jordan, Born, Pauli, Dirac and Schrödinger formulated quantum mechanics, is the topic of the opening chapter of the book.
Rovelli only devotes a short chapter to discuss interpretations in general. This is certainly understandable, since the author’s main target is to discuss his own brainchild: relational quantum mechanics. This approach, however, does not do justice to popular ideas among experts, such as the many-worlds interpretation. The reader may be surprised not to find anything about the Copenhagen (or, more appropriately, Bohr’s) interpretation. This is for very good reason, however, since it is not generally considered to be a coherent interpretation. Having mostly historical significance, it has served as inspiration to approaches that keep the spirit of Bohr’s ideas, like consistent histories (not mentioned in the book at all), or Rovelli’s relational quantum mechanics.
Relational quantum mechanics was introduced by Rovelli in an original technical article in 1996 (Int. J. Theor. Phys.35 1637). Helgoland presents a simplified version of these ideas, explained in more detail in Rovelli’s article, and in a way suitable for a more general audience. The original article, however, can serve as very nice complementary reading for those with some physics background. Relational quantum mechanics claims to be compatible with several of Bohr’s ideas. In some ways it goes back to the original ideas of Heisenberg by formulating the theory without a reference to a wavefunction. The properties of a system are defined only when the system interacts with another system. There is no distinction between observer and observed system. Rovelli meticulously embeds these ideas in a more general historical and philosophical context, which he presents in a captivating manner. He even speculates whether this way of thinking can help us understand topics that, in his opinion, are unrelated to quantum mechanics, such as consciousness.
Helgoland’s potential audience is very diverse and manages to transcend the fact that it is written for the general public. Professionals from both the sciences and the humanities will certainly learn something, especially if they are not acquainted with the nuances of the interpretations of modern physics. The book, however, as is explicitly stated by Rovelli, takes a partisan stance, aiming to promote relational quantum mechanics. As such, it may give a somewhat skewed view of the topic. In that respect, it would be a good idea to read it alongside books with different perspectives, such as Sean Carroll’s Something Deeply Hidden (2019) and Adam Becker’s What is Real? (2018).
The European Consortium for Astroparticle theory (EuCAPT) held its first annual symposium from 5 to 7 May. Hundreds of theoretical physicists from Europe and beyond met online to discuss the present and future of astroparticle physics and cosmology, in a dense and exciting meeting that featured 29 invited presentations, 42 lightning talks by young researchers, and two community-wide brainstorming sessions.
Participants discussed a wide array of topics at the interface between particle physics, astrophysics and cosmology, with particular emphasis on the challenges and opportunities for these fields in the next decade. Rather than focusing on experimental activities and the discoveries they might enable, the sessions were structured around thematic areas and explored the interdisciplinary multi-messenger aspects of each.
Two sessions were dedicated to cosmology, exploring the early and late universe. As stressed by Geraldine Servant (Hamburg), several unresolved puzzles of particle physics – such as the origin of dark matter, the baryon asymmetry, and inflation – are directly linked to the early universe, and new observational probes may soon shed new light on them.
Julien Lesgourgues (Aachen) showed how the very same puzzles are also linked to the late universe, and cautiously elaborated on a series of possible inconsistencies between physical quantities inferred from early- and late-universe probes, for example the Hubble constant. Those inconsistencies represent both a challenge and an extraordinary opportunity for cosmology, as they might “break” the standard Lambda–cold-dark-matter model of cosmology, and allow us to gain insights into the physics of dark matter, dark energy and gravity.
We are witnessing a proliferation of theoretically well-motivated models
New strategies to go beyond the standard models of particle physics and cosmology were also discussed by Marco Cirelli (LPTHE) and Manfred Lindner (Heidelberg), in the framework of dark-matter searches and neutrino physics, respectively. Progress in both fields is currently not limited by a lack of ideas – we are actually witnessing a proliferation of theoretically well-motivated models – but by the difficulty of identifying experimental strategies to conclusively validate or rule them out. Much of the discussion here concerned prospects for detecting new physics with dedicated experiments and multi-messenger observations.
Gravitational waves have added a new observational probe in astroparticle physics and cosmology. Alessandra Buonanno (Max Planck Institute for Gravitational Physics) illustrated the exciting prospects for this new field of research, whose potential for discovering new physics is attracting enormous interest from particle and astroparticle theorists. The connection between cosmic rays, gamma rays and high-energy neutrinos was explored in the final outlook by Elena Amato (Arcetri Astrophysical Observatory), who highlighted how progress in theory and observations is leading the community to reconsider some long-held beliefs – such as the idea that supernova remnants are the acceleration sites of cosmic rays up to the so-called “knee” – and stimulating new ideas.
In line with EuCAPT’s mission, the local organisers and the consortium’s steering committee organised a series of community-building activities. Participants stressed the importance of supporting diversity and inclusivity, a continuing high priority for EuCAPT, while a second brainstorming session was devoted to the discussion of the EuCAPT white paper currently being written, which should be published by September. Last but not least, Hannah Banks (Cambridge), Francesca Capel (TU Munich) and Charles Dalang (University of Geneva) received prizes for the best lightning talks, and Niko Sarcevic (Newcastle) was awarded an “outstanding contributor” prize for the help and support she provides for the analysis of the EuCAPT census (pictured).
The next symposium will take place in 2022, hopefully in person, at CERN.
Throughout the history of nuclear, particle and astroparticle physics, novel detector concepts have paved the way to new insights and new particles, and will continue to do so in the future. To help train the next generation of innovators, noted experimental particle physicists Hermann Kolanoski (Humboldt University Berlin and DESY) and Norbert Wermes (University of Bonn) have written a comprehensive textbook on particle detectors. The authors use their broad experience in collider and underground particle-physics experiments, astroparticle physics experiments and medical-imaging applications to confidently cover the spectrum of experimental methods in impressive detail.
Particle Detectors – Fundamentals and Applications combines in a single volume the syllabus also found in two well-known textbooks covering slightly different aspects of detectors: Techniques for Nuclear and Particle Physics Experiments by W R Leo and Detectors for Particle Radiation by Konrad Kleinknecht. Kolanoski and Wermes’ book supersedes them both by being more up-to-date and comprehensive. It is more detailed than Particle Detectors by Claus Grupen and Boris Shwartz – another excellent and recently published textbook with a similar scope – and will probably attract a slightly more advanced population of physics students and researchers. This new text promises to become a particle-physics analogue of the legendary experimental-nuclear-physics textbook Radiation Detection and Measurement by Glenn Knoll.
The book begins with a comprehensive warm-up chapter on the interaction of charged particles and photons with matter, going well beyond a typical textbook level. This is followed by a very interesting discussion of the transport of charge carriers in media in magnetic and electric fields, and – a welcome novelty – signal formation, using the method of “weighting fields”. The main body of the book is devoted first to gaseous, semiconductor, Cherenkov and transition-radiation detectors, and then to detector systems for tracking, particle identification and calorimetry, and the detection of cosmic rays, neutrinos and exotic matter. Final chapters on electronics readout, triggering and data acquisition complete the picture.
Particle Detectors – Fundamentals and Applications is best considered a reference for lectures on experimental methods in particle and nuclear physics for postgraduate-level students. The book is easy to read, and conceptual discussions are well supported by numerous examples, plots and illustrations of excellent quality. Kolanoski and Wermes have undoubtedly written a gem of a book, with value for any experimental particle physicist, be they a master’s student, PhD student or accomplished researcher looking for detector details outside of their expertise.
Amalia Ballarino of CERN has received the 2021 James Wong Award from the Institute of Electrical and Electronics Engineers (IEEE) for her significant and continuing contributions in the field of superconducting materials. The IEEE citation recognises her for: “leading successful R&D programs that establish a winning role for high temperature and MgB2 superconductors in accelerator applications; piloting the development of MgB2 wire suitable for cabling and its incorporation into a multi-kA power transmission system operating at 25 K, and directing the project to industrialise eight such systems for which over 1000 km of wire have been produced; promoting fruitful cooperation between research and industry; and launching R&D activity based on the use of superconductors (Nb-Ti, Nb3Sn, MgB2 and high-temperature superconductors) for future particle accelerators.
Ballarino was responsible for the several-thousand current leads that power the superconducting magnets of the LHC, including those based on the high-temperature superconductor BSCCO-2223, which have been the first large-scale commercial application of high-temperature superconductors. She was awarded Superconducting Week’s “Superconductor Industry Person of the Year 2006” for the development. Following work on the commissioning of the LHC, Ballarino proposed cold-powering systems that use high-current MgB2 transfer lines for feeding the new superconducting magnets of the High-Luminosity LHC (HL-LHC). She started a collaboration with industry to develop the conductor in the form of wire suitable for cabling. The wire has been successfully delivered to CERN in large quantities, while the cold-powering systems have been developed and qualified and they are now being industrialised.
CERN is home to more winners than any other institution
Ballarino joined CERN as PhD student. She is section leader in CERN’s magnets, superconductors and cryostats group and, as from January 2021, deputy group leader. The IEEE cited her service to the community as lecturer, member of program committees for international conferences, and technical editor and reviewer of papers for scientific journals. “In my opinion, this recognition has been a long time in coming,” says Bruce Strauss, past president and treasurer of the IEEE council on superconductivity.
The IEEE James Wong Award (formally named “Award for Continuing and Significant Contributions in the Field of Applied Superconductivity” until 2013) comes with a $5000 honorarium and a pure-niobium medal. It has been granted annually by the IEEE council on superconductivity since 2000, and CERN is home to more winners than any other institution, with Daniel Leroy, Lucio Rossi, Herman ten Kate, Robert Aymar, Arnaud Devred and Luca Bottura recognised in previous years.
This year’s Future Circular Collider (FCC) Week took place online from 28 June to 2 July, attracting 700 participants from all over the world to debate the next steps needed to produce a feasibility report in 2025/2026, in time for the next update to the European Strategy for Particle Physics in 2026/2027. The current strategy, agreed in 2020, sets an electron–positron Higgs factory as the highest priority facility after the LHC, along with the investigation of the technical and financial feasibility of such a Higgs factory, followed by a high-energy hadron collider placed in the same 100 km tunnel. The FCC feasibility study will focus on the first stage (tunnel and e+e– collider) in the next five years.
Although the FCC is a long-term project with a horizon up to the 22nd century, its timescales are rather tight. A post-LHC collider should be operational around the 2040s, ensuring a smooth continuation from the High-Luminosity LHC, so construction would need to begin in the early 2030s. Placement studies to balance geological and territorial constraints with machine requirements and physics performance suggest that the most suitable scenarios are based on a 92 km-circumference tunnel with eight surface sites.
The next steps are subsurface investigations of high-risk areas, surface-site initial-state analysis and verification of in-principle feasibility with local authorities. A “Mining the Future” competition has been launched to solicit ideas for how to best use the nine million cubic metres of molasse that would be excavated from the tunnel.
The present situation in particle physics is reminiscent of the early days of superconductivity
A highlight of the week was the exploration of the physics case of a post-LHC collider. Matthew Reece (Harvard University) identified dark matter, the baryon asymmetry and the origin of primordial density perturbations as key experimental motivations, and the electroweak hierarchy problem, the strong CP problem and the mystery of flavour mixing patterns as key theoretical motivations. The present situation in particle physics is reminiscent of the early days of superconductivity, he noted, when we had a phenomenological description of symmetry breaking in superconductivity, but no microscopic picture. Constraining the shape of the Higgs potential could allow a similar breakthrough for electroweak symmetry breaking. Regarding recent anomalous measurements, such as those of the muon’s magnetic moment, Reece noted that while these measurements could give us the coefficients of one higher dimension operator in an effective-field-theory description of new physics, only colliders can systematically produce and characterise the nature of any new physics. FCC-ee and FCC-hh both have exciting and complementary roles to play.
A key technology for FCC-ee is the development of efficient superconducting radio-frequency (SRF) cavities to compensate for the 100 MW synchrotron radiation power loss in all modes of operation from the Z pole up to the top threshold at 365 GeV. A staged RF system is foreseen as the baseline scenario, with low-impedance single-cell 400 MHz Nb/Cu cavities for Z running replaced by four-cell Nb/Cu cavities for W and Higgs operation, and later augmented by five-cell 800 MHz bulk Nb cavities at the top threshold.
As well as investigations into the use of HIPIMS coating and the fabrication of copper substrates, an innovative slotted waveguide elliptical (SWELL) cavity design was presented that would operate at 600 or 650 MHz. SWELL cavities optimise the surface area, simplify the coating process and avoid the need for welding in critical areas, which could reduce the performance of the cavity. The design profits from previous work on CLIC, and may offer a simplified installation schedule while also finding applications outside of high-energy physics. A prototype will be tested later this year.
Several talks also pointed out synergies with the RF systems needed for the proposed electron–ion collider at Brookhaven and the powerful energy-recovery linac for experiments (PERLE) project at Orsay, and called for stronger collaboration between the projects.
Machine design
Another key aspect of the study regards the machine design. Since the conceptual design report last year, the pre-injector layout for FCC-ee has been simplified, and key FCC-ee concepts have been demonstrated at Japan’s SuperKEKB collider, including a new world-record luminosity of 3.12 × 1034 cm–2 s–1 in June with a betatron function of βγ* = 1 mm. Separate tests squeezed the beam to just βγ* = 0.8 mm in both rings.
Other studies reported during FCC Week 2021 demonstrated that hosting four experiments is compatible with a new four-fold symmetric ring. This redundancy is thought to be essential for high-precision measurements, and different detector solutions will be invaluable in uncovering hidden systematic biases. The meeting also followed up on the proposal for energy-recovery linacs (ERLs) at FCC-ee, potentially extending the energy reach to 600 GeV if deemed necessary during the previous physics runs. First studies for the use of the FCC-ee booster as a photon source were also presented, potentially leading to applications in medicine and industry, precision QED studies and fundamental-symmetry tests.
Participants also tackled concepts for power reduction and power recycling, to ensure that FCC is sustainable and environmentally friendly. Ideas relating to FCC-ee include making the magnets superconducting rather than normal conducting, improving the klystron efficiency, using ERLs and other energy-storage devices, designing “twin” dipole and quadrupole magnets with a factor-two power saving, and coating SRF cavities with a high-temperature superconductor.
All in all, FCC Week 2021 saw tremendous progress across different areas of the study. The successful completion of the FCC Feasibility Study (2021–2025) will be a crucial milestone for the future of CERN and the field.
Recent years have seen rapid growth in high-energy gamma-ray astronomy, with the first measurement of TeV photons from gamma-ray bursts by the MAGIC telescope and the first detection of gamma rays with energies above 100 TeV by the HAWC observatory.
Now, the Large High Altitude Air Shower Observatory (LHAASO) in China has increased the energy scale at which the universe has been observed by a further order of magnitude. The recent LHAASO detection provides the first clear evidence of the presence of galactic “pevatrons”: sources in the Milky Way capable of accelerating protons and electrons to PeV energies. Although PeV cosmic rays are known to exist, magnetic fields pervading the universe perturb their direction and therefore do not allow their origin to be traced. The gamma rays produced by such cosmic-rays, on the other hand, point directly to their source.
Wide field of view
LHAASO is located in the mountains of the Sichuan province of China and offers a wide field of view to study both high-energy cosmic and gamma rays. Once completed, the observatory will contain a water Cherenkov detector with a total area of about 78,000 m2, 18 widefield- of-view Cherenkov telescopes and a 1 km2 array of more than 5000 scintillator- based electromagnetic detectors (EDs). Finally, more than 1000 underground water Cherenkov tanks (the MDs) are placed over the grid to detect muons.
The latter two detectors, of which only half were finished during data-taking for this study, are used to directly detect the showers produced when high-energy particles interact with the Earth’s atmosphere. The EDs detect the shower profile and incoming angle, using charge and timing information of the detector array, while the MDs are used to distinguish hadronic showers from the electromagnetic showers produced by high-energy gamma rays. Thanks to both its large size and the MDs, LHAASO will ultimately be two orders of magnitude more sensitive at 100 TeV than the HAWC facility in Mexico, the previous most sensitive detector of this type.
The measurements reported by the Chinese-led international LHAASO collaboration reveal a total of 12 sources Astrowatch Mountain observatory nets PeV gamma rays located across the galactic plane (see image above). This distribution is expected, since gamma rays at such energies have a high cross-section for pair production with the cosmic microwave background and therefore the universe starts to become opaque at energies exceeding tens to hundreds of TeV, leaving only sources within our galaxy visible. Of the 12 presented sources, only the Crab nebula can be directly confirmed. This substantiates the pulsar-wind nebulae as a source in which electrons are accelerated beyond PeV energies, which in turn are responsible for the gamma rays through inverse Compton scattering.
Of specific interest is the source responsible for the photon with the highest energy, 1.4 PeV
The origin of the other photons remains unknown as the observed emission regions contain several possible sources within them. The sizes of the emission regions exceed the angular resolution of LHAASO, however, indicating that emission takes place over large scales. Of specific interest is the source responsible for the photon with the highest energy, 1.4 PeV. This came from a region containing both a supernova remnant as well as a star-forming cluster, both of which are prime theoretical candidates for hadronic pevatrons.
Tip of the iceberg
More detailed spectrometry as well as morphological measurements, in which the differences in emission intensity throughout the sources are measured, could allow the sources of > 100 TeV gamma rays to be identified in the next one or two years, say the authors. Furthermore, as the current 12 sources were visible using only one year of data from half the detector, it is clear that LHAASO is only seeing the tip of iceberg when it comes to high-energy gamma rays.
More than a century after its discovery, the proton remains a source of intrigue, its charge-radius and spin posing puzzles that are the focus of intense study. But what of its mortal sibling, the neutron? In recent years, discrepancies between measurements of the neutron lifetime using different methods constitute a puzzle with potential implications for cosmology and particle physics. The neutron lifetime determines the ratio of protons to neutrons at the beginning of big-bang nucleosynthesis and thus affects the yields of light elements, and it is also used to determine the CKM matrix-element Vud in the Standard Model.
The neutron-lifetime puzzle stems from measurements using two techniques. The “bottle” method counts the number of surviving ultra-cold neutrons contained in a trap after a certain period, while the “beam” method uses the decay probability of the neutron obtained from the ratio of the decay rate to an incident neutron flux. Back in the 1990s, the methods were too imprecise to worry about differences between the results. Today, however, the average neutron lifetime measured using the bottle and beam methods, 879.4 ± 0.4 s and 888.0 ± 2.0 s, respectively, stand 8.6 s (or 4σ) apart.
We think it will take two years to obtain a competitive result from our experiment
Kenji Mishima
In an attempt to shed light on the issue, a team at Japan’s KEK laboratory in collaboration with Japanese universities has developed a new experimental setup. Similar to the beam method, it compares the decay rate to the reaction rate of neutrons in a pulsed beam from the Japan Proton Accelerator Research Complex (J-PARC). The decay rate and the reaction rate are determined by simultaneously detecting electrons from the neutron decay and protons from the reaction 3He →3H in a 1 m-long time-projection chamber containing diluted 3He, removing some of the systematic uncertainties that affect previous beam methods. The experiment is still in its early stages, and while the first results have been released – τn = 898 ± 10(stat)+15–18 (sys) s – the uncertainty is currently too large to draw conclusions.
“In the current situation, it is important to verify the puzzle by experiments in which different systematic errors dominate,” says Kenji Mishima of KEK, adding that further improvements in the statistical and systematic uncertainties are underway. “We think it will take two years to obtain a competitive result from our experiment.”
Several new-physics scenarios have been proposed as solutions of the neutron lifetime puzzle. These include exotic decay modes involving undetectable particles with a branching ratio of about 1%, such as “mirror neutrons” or dark-sector particles.
Pixel detectors have their roots in photography. Up until 50 years ago, every camera contained a roll of film on which images were photochemically recorded with each exposure, after which the completed roll was sent to be “developed” to finally produce eagerly awaited prints a week or so later. For decades, film also played a big part in particle tracking, with nuclear emulsions, cloud chambers and bubble chambers. The silicon chip, first unveiled to the world in 1961, was to change this picture forever.
During the past 40 years, silicon sensors have transformed particle tracking in high-energy physics experiments
By the 1970s, new designs of silicon chips were invented that consisted of a 2D array of charge-collection sites or “picture elements” (pixels) below the surface of the silicon. During the exposure time, an image focused on the surface generated electron–hole pairs via the photoelectric effect in the underlying silicon, with the electrons collected as signal information in the pixels. These chips came in two forms: the charge-coupled device (CCD) and the monolithic active pixel sensor (MAPS) – more commonly known commercially as the CMOS image sensor (CIS). Willard Boyle and George Smith of Bell Labs in the US were awarded the Nobel Prize for Physics in 2009 for inventing the CCD.
In a CCD, the charge signals are sequentially transferred to a single on-chip output circuit by applying voltage pulses to the overlying electrode array that defines the pixel structure. At the output circuit the charge is converted to a voltage signal to enable the chip to interface with external circuitry. In the case of the MAPS, each pixel has its own charge-integrating detection circuitry and a voltage signal is again sequentially read out from each by on-chip switching or “scanning” circuitry. Both architectures followed rapid development paths, and within a couple of decades had completely displaced photographic film in cameras.
For the consumer camera market, CCDs had the initial lead, which passed to MAPS by about 1995. For scientific imaging, CCDs are preferred for most astronomical applications (most recently the 3.2 Gpixel optical camera for the Vera Rubin Observatory), while MAPS are the preferred option for fast imaging such as super-resolution microscopy, cryoelectron microscopy and pioneering studies of protein dynamics at X-ray free-electron lasers. Recent CMOS imagers with very small, low-capacitance pixels achieve sufficiently low noise to detect single electrons. A third member of the family is the hybrid pixel detector, which is MAPS-like in that the signals are read out by scanning circuitry, but in which the charges are generated in a separate silicon layer that is connected, pixel by pixel, to a readout integrated circuit (ROIC).
During the past 40 years, these devices (along with their silicon-microstrip counterparts, to be described in a later issue) have transformed particle tracking in high-energy physics experiments. The evolution of these device types is intertwined to such an extent that any attempt at historical accuracy, or who really invented what, would be beyond the capacity of this author, for which I humbly apologise. Space constraints have also led to a focus on the detectors themselves, while ignoring the exciting work in ROIC development, cooling systems, mechanical supports, not to mention the advanced software for device simulation, the simulation of physics performance, and so forth.
CCD design inspiration
The early developments in CCD detectors were disregarded by the particle-detector community. This is because gaseous drift chambers, with a precision of around 100 μm, were thought to be adequate for all tracking applications. However, the 1974 prediction by Gaillard, Lee and Rosner that particles containing charm quarks “might have lifetimes measurable in emulsions”, followed by the discovery of charm in 1975, set the world of particle-physics instrumentation ablaze. Many groups with large budgets tried to develop or upgrade existing types of detectors to meet the challenge: bubble chambers became holographic; drift chambers and streamer chambers were pressurised; silicon microstrips became finer-pitched, etc.
Illustrations of a CCD (left), MAPS (middle) and hybrid chip (right). The first two typically contain 1 k × 1 k pixels, up to 4 k × 4 k or beyond by “stitching”, with an active layer thickness (depleted) of about 20 µm and a highly doped bulk layer back-thinned to around 100 µm, enabling a low-mass tracker, even potentially bent into cylinders round the beampipe.
The CCD (where I is the imaging area, R the readout register, TG the transfer gate, CD the collection diode, and S, D, G the source, drain and gate of the sense transistor) is pixellised in the I direction by conducting gates. Signal charges are shifted in this direction by manipulating the gate voltages so that the image is shifted down, one row at a time. Charges from the bottom row are tipped into the linear readout register, within which they are transferred, all together in the orthogonal direction, towards the output node. As each signal charge reaches the output node, it modulates the voltage on the gate of the output transistor; this is sensed, and transmitted off-chip as an analog signal.
In a MAPS chip, pixellisation is implemented by orthogonal channel stops and signal charges are sensed in-pixel by a tiny front-end transistor. Within a depth of about 1 µm below the surface, each pixel contains complex CMOS electronics. The simplest readout is “rolling shutter”, in which peripheral logic along the chip edge addresses rows in turn, and analogue signals are transmitted by column lines to peripheral logic at the bottom of the imaging area. Unlike in a CCD, the signal charges never move from their “parent” pixel.
In the hybrid chip, like a MAPS, signals are read out by scanning circuitry. However, the charges are generated in a separate silicon layer that is connected, pixel by pixel, to a readout integrated circuit. Bump-bonding interconnection technology is used to keep up with pixel miniaturisation.
The ACCMOR Collaboration (Amsterdam, CERN, Cracow, Munich, Oxford, RAL) had built a powerful multi-particle spectrometer, operating at CERN’s Super Proton Synchrotron, to search for hadronic production of the recently-discovered charm particles, and make the first measurements of their lifetimes. We in the RAL group picked up the idea of CCDs from astronomers at the University of Cambridge, who were beginning to see deeper into space than was possible with photographic film (see left figure in “Pixel architectures” panel). The brilliant CCD developers in David Burt’s team at the EEV Company in Chelmsford (now Teledyne e2v) suggested designs that we could try for particle detection, notably to use epitaxial silicon wafers with an active-layer thickness of about 20 μm. At a collaboration meeting in Cracow in 1978, we demonstrated via simulations that just two postage-stamp-sized CCDs, placed 1 and 2 cm beyond a thin target, could cover the whole spectrometer aperture and might be able to deliver high-quality topological reconstruction of the decays of charm particles with expected lifetimes of around 10–13 s.
We still had to demonstrate that these detectors could be made efficient for particle detection. With a small telescope comprising three CCDs in the T6 beam from CERN’s Proton Synchrotron we established a hit efficiency of more than 99%, a track measurement precision of 4.5 μm in x and y, and two-track resolution of 40 μm. Nothing like this had been seen before in an electronic detector. Downstream of us, in the same week, a Yale group led by Bill Willis obtained signals from a small liquid-argon calorimeter. A bottle of champagne was shared!
It was then a simple step to add two CCDs to the ACCMOR spectrometer and start looking for charm particles. During 1984, on the initial shift, we found our first candidate (see “First charm” figure), which, after adding the information from the downstream microstrips, drift chambers (with two large aperture magnets for momentum measurement), plus a beautiful assembly of Cherenkov hodoscopes from the Munich group, proved to be a D+→ K+π+π– event.
It was more challenging to develop a CCD-based vertex detector for the SLAC Large Detector (SLD) at the SLAC Linear Collider (SLC), which became operational in 1989. The level of background radiation required a 25 mm-radius beam pipe, and the physics demanded large solid-angle coverage, as in all general-purpose collider detectors. The physics case for SLD had been boosted by the discovery in 1983 that the lifetime of particles containing b quarks was longer than for charm, in contrast to the theoretical expectation of being much shorter. So the case for deploying high-quality vertex detectors at SLC and LEP, which were under construction to study Z0 decays, was indeed compelling (see “Vertexing” figure). All four LEP experiments employed a silicon-microstrip vertex detector.
Early in the silicon vertex-detector programme, e2V perfected the art of “stitching” reticles limited to an area of 2 × 2 cm2, to make large CCDs (8 × 1.6 cm2 for SLD). This enabled us to make a high-performance vertex detector that operated from 1996 until SLD shut down in 1998, and which delivered a cornucopia of heavy-flavour physics from Z0 decays (see “Pioneering pixels” figure). During this time, the LEP beam pipe, limited by background to 54 mm radius, permitted its experiments’ microstrip-based vertex detectors to do pioneering b physics. But it had reduced capability for the more elusive charm, which was shorter lived and left fewer decay tracks.
Between LEP with its much higher luminosity and SLD with its small beam pipe, state-of-the-art vertex detector and highly polarised electron beam, the study of Z0 decays yielded rich physics. Highlights included very detailed studies of an enormous sample of gluon jets from Z0→ b b g events, with cleanly tagged b jets at LEP, and Ac, the parity-violation parameter in the coupling of the Z0 to c-quarks, at SLD. However, the most exciting discovery of that era was the top quark at Fermilab, in which the SVX microstrip detector of the CDF detector played an essential part (see “Top detector” figure). This triggered a paradigm shift. Before then, vertex detectors were an “optional extra” in experiments; afterwards, they became obligatory in every energy frontier detector system.
Hybrid devices
While CCDs pioneered the use of silicon pixels for precision tracking, their use was restricted by two serious limitations: poor radiation tolerance and long readout time (tens of ms due to the need to transfer the charge signals pixel by pixel through a single output circuit). There was clearly a need for pixel detectors in more demanding environments, and this led to the development of hybrid pixel detectors. The idea was simple: reduce the strip length of well-developed microstrip technology to equal its width, and you had your pixel sensor. However, microstrip detectors were read out at one end by ASIC (application-specific integrated circuit) chips having their channel pitch matched to that of the strips. For hybrid pixels, the ASIC readout required a front-end circuit for each pixel, resulting in modules with the sensor chip facing the readout chip, with electrical connections made by metal bump-bonds (see right figure in “Pixel architectures” panel). The use of relatively thick sensor layers (compared to CCDs) compensated for the higher node capacitance associated with the hybrid front-end circuit.
Although the idea was simple, its implementation involved a long and challenging programme of engineering at the cutting edge of technology. This had begun by about 1988, when Erik Heijne and colleagues in the CERN microelectronics group had the idea to fit full nuclear-pulse processing electronics in every pixel of the readout chip, with additional circuitry such as digitisation, local memory and pattern recognition on the chip periphery. With a 3 μm feature size, they were obliged to begin with relatively large pixels (75 × 500 μm), and only about 80 transistors per pixel. They initiated the RD19 collaboration, which eventually grew to 150 participants, with many pioneering developments over a decade, leading to successful detectors in at least three experiments: WA97 in the Omega Spectrometer; NA57; and forward tracking in DELPHI. As the RD19 programme developed, the steady reduction in feature size permitted the use of in-pixel discriminators and fast shapers that enhanced the noise performance, even at high rates. This would be essential for operation of large hybrid pixel systems in harsh environments, such as ATLAS and CMS at the LHC. RD19 initiated a programme of radiation hardness by design (enclosed-gate transistors, guard rings, etc), which was further developed and broadly disseminated by the CERN microelectronics group. These design techniques are now used universally across the LHC detector systems. There is still much to be learned, and advances to a smaller feature size bring new opportunities but also surprises and challenges.
The advantages of the hybrid approach include the ability to choose almost any commercial CMOS process and combine it with the sensor best adapted to the application. This can deliver optimal speed of parallel processing, and radiation hardness as good as can be engineered in the two component chips. The disadvantages include a complex and expensive assembly procedure, high power dissipation due to large node capacitance, and more material than is desirable for a tracking system. Thanks to the sustained efforts of many experts, an impressive collection of hybrid pixel tracking detectors has been brought to completion in a number of detector facilities. As vertex detectors, their greatest triumph has been in the inferno at the heart of ATLAS and CMS where, for example, they were key to the recent measurement of the branching ratio for H → b b .
Facing up to the challenge
The high-luminosity upgrade to the LHC (HL-LHC) is placing severe demands on ATLAS and CMS, none more so than developing even more powerful hybrid vertex detectors to accommodate a “pileup” level of 200 events per bunch crossing. For the sensors, a 3D variant invented by Sherwood Parker has adequate radiation hardness, and may provide a more secure option than the traditional planar pixels, but this question is still open. 3D pixels have already proved themselves in ATLAS, for the insertable B layer (IBL), where the signal charge is drifted transversally within the pixel to a narrow column of n-type silicon that runs through the thickness of the sensor. But for HL-LHC, the innermost pixels need to be at least five times smaller in area than the IBL, putting extreme pressure on the readout chip. The RD53 collaboration led by CERN has worked for years on the development of an ASIC using 65 nm feature size, which enables the huge amount of radiation-resistant electronics to fit within the pixel area, reaching the limit of 50 × 50 μm2. Assembling these delicate modules, and dealing with the thermal stresses associated with the power dissipation in the warm ASICs mechanically coupled to the cold sensor chips, is still a challenge. These pixel tracking systems (comprising five layers of barrel and forward trackers) will amount to about 6 Gpixels – seven times larger than before. Beyond the fifth layer, conditions are sufficiently relaxed that microstrip tracking will still be adequate.
The latest experiment to upgrade from strips to pixels is LHCb, which has an impressive track record of b and charm physics. Its adventurous Vertex Locator (VELO) detector has 26 disks along the beamline, equipped with orthogonally oriented r and ϕ microstrips, starting from inside the beampipe about 8 mm from the LHC beam axis. LHCb has collected the world’s largest sample of charmed hadrons, and with the VELO has made a number of world-leading measurements including the discovery of CP violation in charm. LHCb is now statistics-limited for many rare decays and will ramp up its event samples with a major upgrade implemented in two stages (see State-of-the-art-tracking for high luminosities).
For the first upgrade, due to begin operation early next year, the luminosity will increase by a factor of up to five, and the additional pattern recognition challenge will be addressed by a new pixel detector incorporating 55 μm pixels and installed even closer (5.1 mm) to the beam axis. The pixel detector uses evaporative CO2 microchannel cooling to allow operation under vacuum. LHCb will double its efficiency by removing the hardware trigger and reading out the data at the beam-crossing frequency of 40 MHz. The new “VeloPix” readout chip will achieve this with readout speeds of up to 20 Gb/s, and the software trigger will select heavy-flavour events based on full event reconstruction. For the second upgrade, due to begin in about 2032, the luminosity will be increased by a further factor of 7.5, allowing LHCb to eventually accumulate 10 times its current statistics. Under these conditions, there will be, on average, 40 interactions per beam crossing, which the collaboration plans to resolve by enhanced timing precision (around 20 ps) in the VELO pixels. The upgrade will require both an enhanced sensor and readout chip. This is an adventurous long-term R&D programme, and LHCb retain a fallback option with timing layers downstream of the VELO, if required.
Monolithic active pixels
Being monolithic, the architecture of MAPS is very similar to that of CCDs (see middle figure in “Pixel architectures” panel). The fundamental difference is that in a CCD, the signal charge is transported physically through some centimetres of silicon to a single charge-sensing circuit in the corner of the chip, while in a MAPS the communication between the signal charge and the outside world is via in-pixel electronics, with metal tracks to the edge of the chip. The MAPS architecture looked very promising from the beginning, as a route to solving the problems of both CCDs and hybrid pixels. With respect to CCDs, the radiation tolerance could be greatly increased by sensing the signal charge within its own pixel, instead of transporting it over thousands of pixels. The readout speed could also be dramatically increased by in-pixel amplitude discrimination, followed by sparse readout of only the hit pixels. With respect to hybrid pixel modules, the expense and complications of bump-bonded assemblies could be eliminated, and the tiny node capacitance opened the possibility of much thinner active layers than were needed with hybrids.
MAPS have emerged as an attractive option for a number of future tracking systems. They offer small pixels where needed (notably for inner-layer vertex detectors) and thin layers throughout the detector volume, thereby minimising multiple scattering and photon conversion, both in barrels and endcaps. Excess material in the forward region of tracking systems such as time-projection and drift chambers, with their heavy endplate structures, has in the past led to poor track reconstruction efficiency, loss of tracks due to secondary interactions, and excess photon conversions. In colliders at the energy frontier (whether pp or e+e–), however, interesting events for physics are often multi-jet, so there are nearly always one or more jets in the forward region.
The first MAPS devices contained little more than a collection diode, a front-end transistor operated as a source follower, reset transistor and addressing logic. They needed only relaxed charge-collection time, so diffusive collection sufficed. Sherwood Parker’s group demonstrated their capability for particle tracking in 1991, with devices processed in the Centre for Integrated Studies at Stanford, operating in a Fermilab test beam. In the decades since, advances in the density of CMOS digital electronics have enabled designers to pack more and more electronics into each pixel. For fast operation, the active volume below the collection diode needs to be depleted, including in the corners of the pixels, to avoid loss of tracking efficiency.
The Strasbourg group led by Marc Winter has a long and distinguished record of MAPS development. As well as highly appreciated telescopes in test beams at DESY for general use, the group supplied its MIMOSA-28 devices for the first MAPS-based vertex detector: a 356 Mpixel two-layer barrel system for the STAR experiment at Brookhaven’s Relativistic Heavy Ion Collider. Operational for a three-year physics run starting in 2014, this detector enhanced the capability to look into the quark–gluon plasma, the extremely hot form of matter that characterised the birth of the universe.
Advances in the density of CMOS digital electronics have enabled designers to pack more and more electronics into each pixel
An ingenious MAPS variant developed by the Semiconductor Laboratory of the Max Planck Society – the Depleted P-channel FET (DEPFET) – is also serving as a high-performance vertex detector in the Belle II detector at SuperKEKB in Japan, part of which is already operating. In the DEPFET, the signal charge drifts to a “virtual gate” located in a buried channel deeper than the current flowing in the sense transistor. As Belle II pushes to even higher luminosity, it is not yet clear which technology will deliver the required radiation hardness.
The small collection electrode of the standard MAPS pixel presents a challenge in terms of radiation hardness, since it is not easy to preserve full depletion after high levels of bulk damage. An important initiative to overcome this was initiated in 2007 by Ivan Perić of KIT, in which the collection electrode is expanded to cover most of the pixel area, below the level of the CMOS electronics, so the charge-collection path is much reduced. Impressive further developments have been made by groups at Bonn University and elsewhere. This approach has achieved high radiation resistance with the ATLASpix prototypes, for instance. However, the standard MAPS approach with small collection electrode may be tunable to achieve the required radiation resistance, while preserving the advantages of superior noise performance due to the much lower sensor capacitance. Both approaches have strong backing from talented design groups, but the eventual outcome is unclear.
Advanced MAPS
Advanced MAPS devices were proposed for detectors at the International Linear Collider (ILC). In 2008 Konstantin Stefanov of the Open University suggested that MAPS chips could provide an overall tracking system of about 30 Gpixels with performance far beyond the baseline options at the time, which were silicon microstrips and a gaseous time-projection chamber. This development was shelved due to delays to the ILC, but the dream has become a reality in the MAPS-based tracking system for the ALICE detector at the LHC, which builds on the impressive ALPIDE chip development by Walter Snoeys and his collaborators. The ALICE ITS-2 system, with 12.5 Gpixels, sets the record for any pixel system (see ALICE tracks new territories). This beautiful tracker has operated smoothly on cosmic rays and is now being installed in the overall ALICE detector. The group is already pushing to upgrade the three central layers using wafer-scale stitching and curved sensors to significantly reduce the material budget. At the 2021 International Workshop on Future Linear Colliders held in March, the SiD concept group announced that they will switch to a MAPS-based tracking system. R&D for vertexing at the ILC is also being revived, including the possibility of CCDs making a comeback with advanced designs from the KEK group led by Yasuhiro Sugimoto.
The most ambitious goal for MAPS-based detectors is for the inner-layer barrels at ATLAS and CMS, during the second phase of the HL-LHC era, where smaller pixels would provide important advantages for physics. At the start of high-luminosity operation, these layers will be equipped with hybrid pixels of 25 × 100 μm2 and 150 μm active thickness, the pixel area being limited by the readout chip, which is based on a 65 nm technology node. Encouraging work led by the CERN ATLAS and microelectronics groups and the Bonn group is underway, and could result in a MAPS option of 25 × 25 μm2, requiring an active-layer thickness of only about 20 μm, using a 28 nm technology node. The improvement in tracking precision could be accompanied by a substantial reduction in power dissipation. The four-times greater pixel density would be more than offset by the reduction in operating voltage, plus the much smaller node capacitance. This route could provide greatly enhanced vertex detector performance at a time when the hybrid detectors will be coming to the end of their lives due to radiation damage. However, this is not yet guaranteed, and an evolution to stacked devices may be necessary. A great advantage of moving to monolithic or stacked devices is that the complex processes are then in the hands of commercial foundries that routinely turn out thousands of 12 inch wafers per week.
High-speed and stacked
During HL-LHC operations there is a need for ultra-fast tracking devices to ameliorate the pileup problems in ATLAS, CMS and LHCb. Designs with a timing precision of tens of picoseconds are advancing rapidly – initially low-gain avalanche diodes, pioneered by groups from Torino, Barcelona and UCSC, followed by other ultra-fast silicon pixel devices. There is a growing list of applications for these devices. For example, ATLAS will have a layer adjacent to the electromagnetic calorimeter in the forward region, where the pileup problems will be severe, and where coarse granularity (~1 mm pixels) is sufficient. LHCb is more ambitious for its stage-two upgrade, as already mentioned. There are several experiments in which such detectors have potential for particle identification, notably π/K separation by time-of-flight up to a momentum limit that depends on the scale of the tracking system, typically 8 GeV/c.
Monolithic and hybrid pixel detectors answer many of the needs for particle tracking systems now and in the future. But there remain challenges, for example the innermost layers at ATLAS and CMS. In order to deliver the required vertexing capability for efficient, cleanly separated b and charm identification, we need pixels of dimensions about 25 × 25 μm, four times below the current goals for HL-LHC. They should also be thinner, down to say 20 μm, to preserve precision for oblique tracks.
Solutions to these problems, and similar challenges in the much bigger market of X-ray imaging, are coming into view with stacked devices, in which layers of CMOS-processed silicon are stacked and interconnected. The processing technique, in which wafers are bonded face-to-face, with electrical contacts made by direct-bond interconnects and through-silicon vias, is now a mature technology and is in the hands of leading companies such as Sony and Samsung. The CMOS imaging chips for phone cameras must be one of the most spectacular examples of modern engineering (see “Up close” figure).
Commercial CMOS image sensor development is a major growth area, with approximately 3000 patents per year. In future these developers, advancing to smaller-node chips, will add artificial intelligence, for example to take a number of frames of fast-moving subjects and deliver the best one to the user. Imagers under development for the automotive industry include those that will operate in the short-wavelength infrared region, where silicon is still sensitive. In this region, rain and fog are transparent, so a driverless car equipped with the technology will be able to travel effortlessly in the worst weather conditions.
While we developers of pixel imagers for science have not kept up with the evolution of stacked devices, several academic groups have over the past 15 years taken brave initiatives in this direction, most impressively a Fermilab/BNL collaboration led by Ron Lipton, Ray Yarema and Grzegorz Deptuch. This work was done before the technical requirements could be serviced by a single technology node, so they had to work with a variety of pioneering companies in concert with excellent in-house facilities. Their achievements culminated in three working prototypes, two for particle tracking and one for X-ray imaging, namely a beautiful three-tier stack comprising a thick sensor (for efficient X-ray detection), an analogue tier and a digital tier (see “Stacking for physics” figure).
The relatively recent term “technology node” embraces a number of aspects of commercial integrated circuit (IC) production. First and foremost is the feature size, which originally meant the minimum line width that could be produced by photolithography, for example the length of a transistor gate. With the introduction of novel transistor designs (notably the FinFET), this term has been generalised to indicate the functional density of transistors that is achievable. At the start of the silicon-tracker story, in the late 1970s, the feature size was about 3 µm. The current state-of-the-art is 5 nm, and the downward Moore’s law trend is continuing steadily, although such narrow lines would of course be far beyond the reach of photolithography. There are other aspects of ICs that are included in the description of any technology node. One is whether they support stitching, which means the production of larger chips by step-and-repeat of reticles, enabling the production of single devices of sizes 10 × 10 cm2 and beyond, in principle up to the wafer scale (which these days is a diameter of 200 or 300 mm, evolving soon to 450 mm). Another is whether they support wafer stacking, which is the production of multi-layer sandwiches of thinned devices using various interconnect technologies such as through-silicon vias and direct-bond interconnects. A third aspect is whether they can be used for imaging devices, which implies optimised control of dark current and noise. For particle tracking, the most advanced technology nodes are unaffordable (the development cost of a single 5 nm ASIC is typically about $500 million, so it needs a large market). However, other features that are desirable and becoming essential for our needs (imaging capability, stitching and stacking) are widely available and less expensive. For example, Global Foundries, which produces 3.5 million wafers per annum, offers these capabilities at their 32 and 14 nm nodes.
For the HL-LHC inner layers, one could imagine a stacked chip comprising a thin sensor layer (with excellent noise performance enabled by an on-chip front-end circuit for each pixel), followed by one or more logic layers. Depending on the technology node, one should be able to fit all the logic (building on the functionality of the RD53 chip) in one or two layers of 25 × 25 μm pixels. The overall thickness could be 20 μm for the imaging layer, and 6 μm per logic layer, with a bottom layer sufficiently thick (~100 μm) to give the necessary mechanical stability to the relatively large stitched chips. The resulting device would still be thin enough for a high-quality vertex detector, and the thin planar sensor-layer pixels including front-end electronics would be amenable to full depletion up to the 10-year HL-LHC radiation dose.
There are groups in Japan (at KEK led by Yasuo Arai, and at RIKEN led by Takaki Hatsui) that have excellent track records for developing silicon-on-insulator devices for particle tracking and for X-ray detection, respectively. The RIKEN group is now believed to be collaborating with Sony to develop stacked devices for X-ray imaging. Given Sony’s impressive achievements in visible-light imaging, this promises to be extremely interesting. There are many applications (for example at ITER) where radiation-resistant X-ray imaging will be of crucial importance, so this is an area in which stacked devices may well own the future.
Outlook
The story of frontier pixel detectors is a bit like that of an art form – say cubism. With well-defined beginnings 50 years ago, it has blossomed into a vast array of beautiful creations. The international community of designers see few boundaries to their art, being sustained by the availability of stitched devices to cover large-area tracking systems, and moving into the third dimension to create the most advanced pixels, which are obligatory for some exciting physics goals.
Face-to-face wafer bonding is now a commercially mature technology
Just like the attribute of vision in the natural world, which started as a microscopic light-sensitive spot on the surface of a unicellular protozoan, and eventually reached one of its many pinnacles in the eye of an eagle, with its amazing “stacked” data processing behind the retina, silicon pixel devices are guaranteed to continue evolving to meet the diverse needs of science and technology. Will they one day be swept away, like photographic film or bubble chambers? This seems unthinkable at present, but history shows there’s always room for a new idea.
In the early 1950s, particle accelerators were national-level activities. It soon became obvious that to advance the field further demanded machines beyond the capabilities of single countries. CERN marked a phase transition in this respect, enabling physicists to cooperate around the development of one big facility. Climate science stands to similarly benefit from a change in its topology.
Modern climate models were developed in the 1960s, but there weren’t any clear applications or policy objectives at that time. Today we need hard numbers about how the climate is changing, and an ability to seamlessly link these changes to applications – a planetary information system for assessing hazards, planning food security, aiding global commerce, guiding infrastructural investments, and much more. National centres for climate modelling exist in many countries. But we need a centre “on steroids”: a dedicated exascale computing facility organised on a similar basis to CERN that would allow the necessary leap in realism.
Quantifying climate
To be computationally manageable, existing climate models solve equations for quantities that are first aggregated over large spatial and temporal scales. This blurs their relationship to physical laws, to phenomena we can measure, and to the impacts of a changing climate on infrastructure. Clouds, for example, are creatures of circulation, particularly vertical air currents. Existing models attempt to infer what these air currents would be given information about much larger scale 2D motion fields. There is a necessary degree of abstraction, which leads to less useful results. We don’t know if air is going up or down an individual mountain, for instance, because we don’t have individual mountains in the model, at best mountain ranges.
In addition to more physical models, we also need a much better quantification of model uncertainty. At present this is estimated by comparing solutions across many low-resolution models, or by perturbing parameters of a given low-resolution model. The particle-physics analogy might be that everyone runs their own low-energy accelerators hoping that coordinated experiments will provide high-energy insights. Concentrating efforts on a few high-resolution climate models, where uncertainty is encoded through stochastic mathematics, is a high-energy effort. It would result in better and more useful models, and open the door to cooperative efforts to systematically explore the structural stability of the climate system and its implications for future climate projections.
Working out climate-science’s version of the Standard Model thus provides the intellectual underpinnings for a “CERN for climate change”. One can and should argue about the exact form such a centre should take, whether it be a single facility or a federation of campuses, and on the relative weight it gives to particular questions. What is important is that it creates a framework for European climate, computer and computational scientists to cooperate, also with application communities, in ways that deliver the maximum benefit for society.
Building momentum
A number of us have been arguing for such a facility for more than a decade. The idea seems to be catching on, less for the eloquence of our arguments, more for the promise of exascale computing. A facility to accelerate climate research in developing and developed countries alike has emerged as a core element of one of 12 briefing documents prepared by the Royal Society in advance of the United Nations Climate Change Conference, COP26, in November. This briefing flanks the European Union’s “Destination Earth” project, which is part of its Green Deal programme – a €1 billion effort over 10 years that envisions the development of improved high-resolution models with better quantified uncertainty. If not anchored in a sustainable organisational concept, however, this risks throwing money to the wind.
Giving a concrete form to such a facility still faces internal hurdles, possibly similar to those faced by CERN in its early days. For example, there are concerns that it will take away funding from existing centres. We believe, and CERN’s own experience shows, that the opposite is more likely true. A “CERN for climate change” would advance the frontiers of the science, freeing researchers to turn their attention to new questions, rather than maintaining old models, and provide an engine for European innovation that extends far beyond climate change.
One day, around the time I started properly reading, somebody gave me a book about the sky, and I found it fascinating to think about what’s beyond the clouds and beyond where the planes and the birds fly. I didn’t know that you could actually make a living doing this kind of thing. At that age, you don’t know what a cosmologist is, unless you happen to meet one and ask what they do. You are just fascinated by questions like “how does it work?” and “how do you know?”.
Was there a point at which you decided to focus on theory?
Not really, and I still think I’m somewhat in-between, in the sense that I like to interpret data and am plugged-in to observational collaborations. I try to make connections to what the data mean in light of theory. You could say that I am a theoretical experimentalist. I made a point to actually go and serve at a telescope a couple of times, but you wouldn’t want to trust me in handling all of the nitty-gritty detail, or to move the instrument around.
What are your research interests?
I have several different research projects, spanning large-scale structure, dark energy, inflation and the cosmic microwave background. But there is a common philosophy: I like to ask how much can we learn about the universe in a way that is as robust as possible, where robust means as close as possible to the truth, even if we have to accept large error bars. In cosmology, everything we interpret is always in light of a theory, and theories are always at some level “spherical cows” – they are approximations. So, imagine we are missing something: how do I know I am missing it? It sounds vague, but I think the field of cosmology is ready to ask these questions because we are swimming in data, drowning in data, or soon will be, and the statistical error bars are shrinking.
This explains your current interest in the Hubble constant. What do you define as the Hubble tension?
Yes, indeed. When I was a PhD student, knowing the Hubble constant at the 40–50% level was great. Now, we are declaring a crisis in cosmology because there is a discrepancy at the very-few-percent level. The Hubble tension is certainly one of the most intriguing problems in cosmology today. Local measurements of the current expansion rate of the universe, for example based on supernovae as standard candles, which do not rely heavily on assumptions about cosmological models, give values that cluster around 73 km s–1 Mpc–1. Then there is another, indirect route to measuring what we believe is the same quantity but only within a model, the lambda-cold-dark-matter (ΛCDM) model, which is looking at the baby universe via the cosmic microwave background (CMB). When we look at the CMB, we don’t measure recession velocities, but we interpret a parameter within the model as the expansion rate of the universe. The ΛCDM model is extremely successful, but the value of the Hubble constant using this method comes out at around 67 km s–1 Mpc–1, and the discrepancy with local measurements is now 4σ or more.
What are the implications if this tension cannot be explained by systematic errors or some other misunderstanding of the data?
The Hubble constant is the only cosmological parameter in the ΛCDM universe that can be measured both directly locally and from classical cosmological observations such as the CMB, baryon acoustic oscillations, supernovae and big-bang nucleosynthesis. It’s also easy to understand what it is, and the error bars are becoming small enough that it is really becoming make-or-break for the ΛCDM model. The Hubble tension made everybody wake up. But before we throw the model out of the window, we need something more.
How much faith do you put in the ΛCDM model compared to, say, the Standard Model of particle physics?
It is a model that has only six parameters, most constrained at the percent level, which explains most of the observations that we have of the universe. In the case of Λ, which quantifies what we call dark energy, we have many orders of magnitude between theory and experiment to understand, and for dark matter we are yet to find a candidate particle. Otherwise, it does connect to fundamental physics and has been extremely successful. For 20 years we have been riding a wave of confirmation of the ΛCDM model, so we need to ask ourselves: if we are going to throw it out, what do we substitute it with? The first thing is to take small steps away from the model, say by adding one parameter. For a while, you could say that maybe there is something like an effective neutrino species that might fix it, but a solution like this doesn’t quite fit the CMB data any more. I think the community may be split 50/50 between being almost ready to throw the model out and keeping working with it, because we have nothing better to use.
It is really becoming make-or-break for the ΛCDM model
Could it be that general relativity (GR) needs to be modified?
Perhaps, but where do we modify it? People have tried to tweak GR at early times, but it messes around with the observations and creates a bigger problem than we already have. So, let’s say we modify in middle times – we still need it to describe the shape of the expansion history of the universe, which is close to ΛCDM. Or we could modify it locally. We’ve tested GR at the solar-system scale, and the accuracy of GPS is a vivid illustration of its effectiveness at a planetary scale. So, we’d need to modify it very close to where we are, and I don’t know if there are modifications on the market that pass all of the observational tests. It could also be that the cosmological constant changes value as the universe evolves, in which case the form of the expansion history would not be the one of ΛCDM. There is some wiggle room here, but changing Λ within the error bars is not enough to fix the mismatch. Basically, there is such a good agreement between the ΛCDM model and the observations that you can only tinker so much. We’ve tried to put “epicycles” everywhere we could, and so far we haven’t found anything that actually fixes it.
What about possible sources of experimental error?
Systematics are always unknowns that may be there, but the level of sophistication of the analyses suggests that if there was something major then it would have come up. People do a lot of internal consistency checks; therefore, it is becoming increasingly unlikely that it is only due to dumb systematics. The big change over the past two years or so is that you typically now have different data sets that give you the same answer. It doesn’t mean that both can’t be wrong, but it becomes increasingly unlikely. For a while people were saying maybe there is a problem with the CMB data, but now we have removed those data out of the equation completely and there are different lines of evidence that give a local value hovering around 73 km s–1 Mpc–1, although it’s true that the truly independent ones are in the range 70–73 km s–1 Mpc–1. A lot of the data for local measurements have been made public, and although it’s not a very glamorous job to take someone else’s data and re-do the analysis, it’s very important.
Is there a way to categorise the very large number of models vying to explain the Hubble tension?
Until very recently, there was an interpretation of early versus late models. But if this is really the case, then the tension should show up in other observables, specifically the matter density and age of the universe, because it’s a very constrained system. Perhaps there is some global solution, so a little change here and a little in the middle, and a little there … and everything would come together. But that would be rather unsatisfactory because you can’t point your finger at what the problem was. Or maybe it’s something very, very local – then it is not a question of cosmology, but whether the value of the Hubble constant we measure here is not a global value. I don’t know how to choose between these possibilities, but the way the observations are going makes me wonder if I should start thinking in that direction. I am trying to be as model agnostic as possible. Firstly, there are many other people that are thinking in terms of models and they are doing a wonderful job. Secondly, I don’t want to be biased. Instead I am trying to see if I can think one-step removed, which is very difficult, from a particular model or parameterisation.
What are the prospects for more precise measurements?
For the CMB, we have the CMB-S4 proposal and the Simons Array. These experiments won’t make a huge difference to the precision of the primary temperature-fluctuation measurements, but will be useful to disentangle possible solutions that have been proposed because they will focus on the polarisation of the CMB photons. As for the local measurements, the Dark Energy Spectroscopic Instrument, which started observations in May, will measure baryon acoustic oscillations at the level of galaxies to further nail down the expansion history of the low-redshift universe. However, it will not help at the level of local measurements, which are being pursued instead by the SH0ES collaboration. There is also another programme in Chicago focusing on the so-called tip of the red-giant-branch technique, with more results to come out. Observations of multiple images from strong gravitational lensing is another promising avenue that is very actively pursued, and, if we are lucky, gravitational waves with optical counterparts will bring in another important piece of the puzzle.
If we are lucky, gravitational waves with optical counterparts will bring in another important piece of
the puzzle
How do we measure the Hubble constant from gravitational waves?
It’s a beautiful measurement, as you can get a distance measurement without having to build a cosmic distance ladder, which is the case with the other local measurements that build distances via Cepheids, supernovae, etc. The recession velocity of the GW source comes from the optical counterpart and its redshift. The detection of the GW170817 event enabled researchers to estimate the Hubble constant to be 70 km s–1 Mpc–1, for example, but the uncertainties using this novel method are still very large, in the region of 10%. A particular source of uncertainty comes from the orientation of the gravitational-wave source with respect to Earth, but this will come down as the number of events increases. So this route provides a completely different window on the Hubble tension. Gravitational waves have been dubbed, rather poetically, “standard sirens”. When these determinations of the Hubble constant become competitive with existing measurements really depends on how many events are out there. Upgrades to LIGO, VIRGO, plus next-generation gravitational-wave observatories will help in this regard, but what if the measurements end up clustering between or beyond the late- and early-time measurements? Then we really have to scratch our heads!
How can results from particle physics help?
Principally, if we learn something about dark matter it could force us to reconsider our entire way to fit the observations, perhaps in a way that we haven’t thought of because dark matter may be hot rather than cold, or something else that interacts in completely different ways. Neutrinos are another possibility. There are models where neutrinos don’t behave like the Standard Model yet still fit the CMB observations. Before the Hubble tension came along, the hope was to say that we have this wonderful model of cosmology that fits really well and implies that we live in a maximally boring universe. Then we could have used that to eventually make the connection to particle physics, say, by constraining neutrino masses or the temperature of dark matter. But if we don’t live in a maximally boring universe, we have to be careful about playing this game because the universe could be much, much more interesting than we assumed.
To provide the best experiences, we use technologies like cookies to store and/or access device information. Consenting to these technologies will allow us to process data such as browsing behavior or unique IDs on this site. Not consenting or withdrawing consent, may adversely affect certain features and functions.
Functional
Always active
The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.
Preferences
The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.
Statistics
The technical storage or access that is used exclusively for statistical purposes.The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.
Marketing
The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.