Ongoing research

Physics analysis of pp collisions

The ATLAS experiment collects pp collisions from the Large Hadron Collider and with the remnants of the collisions we can measure properties of elementary particles and search for completely new phenomena. The ATLAS group in Uppsala follows three main analysis tracks: searches for pair-production of the Higgs boson, trying to find out if the Higgs boson is an elementary particle, and using top quarks and Higgs bosons to look for dark matter, as explained in the following.

Higgs boson pair production (HH)

Contact person: Arnaud Ferrari

The Standard Model predicts the existence of a scalar particle, the Higgs boson, discovered by ATLAS and CMS at the LHC in 2012. But it also predicts that this particle has the unique feature to couple to itself. For a complete measurement of the so-called Higgs potential and thereby a full validation of the Standard Model, events must be observed, in which two 125 GeV Higgs bosons are produced simultaneously. Besides the Higgs self-coupling (illustrated in Figure b), another process leading to two Higgs bosons is via a box of heavy particles, mostly top quarks (see Figure a).

Unfortunately, these diagrams interfere destructively and the cross-section for Higgs boson pair production at the LHC is very small, three orders of magnitude less than single Higgs boson production. Hence, one does not expect to have evidence for this process before the high luminosity upgrade of LHC (HL-LHC). On the other hand, if Higgs boson pairs were to be observed in the current data or during Run-3 of the LHC, this would indicate new physics. Modifications of the Higgs boson self-coupling with respect to the Standard Model prediction may enhance the production cross-section by modifying the interference pattern. Alternatively, Higgs boson pairs could be produced in the decay of a new resonance, as illustrated in Figure c.

Higgs boson pair decays

The ATLAS group in Uppsala is pursuing searches for Higgs boson pairs in the final state with two b-quarks and two tau-leptons. With 36 fb–1 of data collected in 2015–2016, this channel provides the most stringent exclusion limit, at about 10 times the Standard Model prediction, among all the individual data analyses performed at the LHC. This search was recently re-interpreted in a scenario where pairs of lepto-quarks are produced, which each decay into a b-quark and a tau-lepton. Our research group also contributes to the statistical combination of ATLAS searches for Higgs boson pairs and to their interpretation in term of the Higgs self-coupling and new physics in the Higgs sector beyond the Standard Model. One of the group members is also the ATLAS convener in the HH sub-group of the LHC Higgs Cross-Section Working Group, in which theoretical models involving Higgs boson pairs are reviewed and recommendations on cross-sections and new benchmarks are gathered.

Solving the Higgs Fine-tuning problem with Top partners (SHIFT)

Contact: Elin Bergeås Kuutmann

The Higgs boson explains why the other elementary particles have masses, but its own mass is an enigma. The theoretical calculation involves quantum corrections which need to cancel at a precision of 1 to 1019 in order to arrive at the measured Higgs boson mass of about 125 GeV (approximately 125 times the mass of the proton). This is called the Higgs fine-tuning problem. Popular proposed solutions include the existence of new heavy particles “top partners”, from supersymmetric models, to stabilise the corrections. The top partners resemble the heaviest known elementary particle, the top quark, hence the name. In Uppsala, we are interested in an alternative approach, namely that the Higgs boson is not an elementary particle but instead composite. This could among other things give rise to top partners in the form of vector like quarks with interesting decay patterns.

The Uppsala team is part of the KAW-funded SHIFT collaboration which also has members from Stockholm University and Chalmers University as well as international collaborators.

Top quarks, Higgs bosons, and Dark Matter (Top/Higgs/DM)

Contact person: Rebeca Gonzalez Suarez

The only place where the Higgs boson can be studied is the LHC, where it was discovered in 2012. There is still much to learn about it, and therefore extensive exploration of the Higgs sector will be done in Run-3, and later on at the HL-LHC.

Top quarks, discovered at the Tevatron in 1995, are produced at an unprecedented rate at the LHC, that is also known as a “top factory”. There are so many top quarks that they are background to virtually any process you want to measure at the LHC. But, as the most massive elementary particles known, they are also quite intriguing and the precise measurement of their properties is a central part of the physics programs of the LHC experiments.

The Higgs boson couples to mass, and therefore, the interplay between the top quark and the Higgs boson is among the most interesting things to study in the next years. Especially in the absence of clear new physics effects in the data. The top-Higgs coupling could hide answers to some of the open questions in particle physics.

One of the most pressing of such questions is the nature of Dark Matter. Dark Matter is prevalent in the Universe but other than that, we know very little about it. Since Dark Matter is affected by gravitational effects, we think it could have mass. And if it has mass, then it may be related to the Higgs boson, and could have something to do with top quarks.

One of the research lines of the ATLAS group in Uppsala University goes into the top-Higgs sector to explore it as a portal of new physics, particularly Dark Matter production at the LHC.

Past activities on charged Higgs bosons (H+)

Contacts: Arnaud Ferrari, Elin Bergeås Kuutmann

In the Standard Model, one complex scalar Higgs doublet is responsible for the electroweak gauge symmetry breaking, giving rise to only one physical Higgs boson. In so-called Two Higgs Doublet Models (2HDM), such as the Minimal Supersymmetric extension of the Standard Model (MSSM), two complex scalar Higgs doublets are responsible for the electroweak symmetry breaking. As a result, the Higgs spectrum consists of five physical states: two charged and three neutral bosons. Until recently, the ATLAS group in Uppsala played a significant role in searching for charged Higgs bosons.

If the charged Higgs boson is lighter than about 170 GeV, it can be produced in the decay of top quarks, which are being copiously pair produced in pp collisions at the LHC. In that case, the main decay is H+→τν, where the tau-lepton may decay either to lighter leptons or hadrons. This decay channel may still have a sizeable branching fraction for heavier charged Higgs bosons, however the dominant decay channel is then H+→tb. Both decay channels have been searched for in ATLAS during the Run-1 and the first half of the Run-2 of LHC, across a broad range of charged Higgs boson masses. Until recently, the group in Uppsala played a significant role in these searches.

ATLAS trigger performance and monitoring

In the LHC, proton collisions happen every 25 nanosecond when the accelerator is operating at maximum capacity. The results of each collision generate about 1 MB of data in the ATLAS detector. Only a small fraction of the events can be kept from sheer limitation in storage and read-out capacity. The ATLAS trigger is the hardware and software based system which takes a first look at each event and decides if it is potentially interesting and should be stored or if it should be deleted forever.

Trigger data quality monitoring

Contact: Elin Bergeås Kuutmann

The ATLAS detector does not only collect data, it should collect the best data. During data-taking it is crucial to continuously monitor the data quality so that problems can be immediately corrected. Otherwise the data collected might be biased or bad in some other way.

At Uppsala University, we are involved in the monitoring of the tau trigger data quality including the maintenance of the software as well as the overall data quality monitoring from a trigger point of view.

The luminosity upgrade of the LHC to a High Luminosity LHC (HL-LHC)

LHC is preparing to increase its luminosity by an order of magnitude in the middle of next decade. The aim is to discover new physics processes and to study known interactions with high precision. This requires a total rebuild of the ATLAS experiment. The most important rebuild is the replacement of the tracker.

Inner TracKer (ITK)

Contact: Richard Brenner, Myrto Asimakopoulou

The tracker is the detector system in ATLAS closest to the point where the protons collide.

The particles created in the proton–proton collisions travel through the tracker. Charged particles deposit small signals in the detector layers of the tracker while passing through. The signal is collected using the tracker’s millions of readout channels. This information can be used to make out the path of the particle. Since the tracker is placed in a 2 T magnetic field, the path will be curved and will allow us to measure the track’s momentum. The spatial resolution of the tracker’s detecting elements is high, and it is possible to identify the decay vertices of a particle with high 3D precision. This allows us to identify particles by their lifetime.

The high interaction rate in ATLAS generates intense radiation that damages the detector elements. A few years into the next decade the radiation damage is already severe and degrades the performance of the tracker. For this reason we are now preparing for the replacement of the current tracker. Uppsala together with our Scandinavian collaborators in Lund, Copenhagen and Oslo are setting up a production line for new silicon strip detector modules that will be more radiation hard, faster and better performing than the predecessors.

In order to use modern production facilities we collaborate with a Swedish electronics industry. Initial small scale R&D is done in-house, targeting a production method that can be ported to industry. Specialized interconnection methods, quality assurance and control will be done in academia.

A detector module for the End-Cap strip part of ITK is shown below.

Detector module

Following the module presented in the figure, the module consists of a large segmented silicon sensor wire bonded to 12 Application Specific Integrated Circuits (ASIC). They read out the minimal signal from the sensor and buffer the information until receiving a trigger(order) to send the data to the Data Acquisition System (DAC). There are two small ASICs in the module to take care of the data transfer between readout ASICs and the DAC. All ASICs are mounted on a low weight, high density flexible Printed Circuit Board (PCB). The final ITK will have a large number of detector modules of various sizes depending on the position in the tracker. Because of the size and complexity of the tracker the detector modules will be made in an international collaboration between Europe, America and Asia. The Scandinavian production line is focused on two module types.

Hardware Tracking for the Trigger

Contact: Richard Brenner, Mikael Mårtenson

A major drawback of a high granularity tracker is the massive amount of data recorded. We are today having a serious bottleneck in the bandwidth for transferring that data from the tracker to the DAQ. In current data recording we can only read out tracking information for 1 physics event in 400. The data bandwidth is however not the only problem. If all tracking data would be read out then we would still not have the capacity to reconstruct that data.

The first problem could easily be solved by adding more services for data transfer. The price would be a significant increase in material in the tracker which would not only degrade the physics performance but also increase the radiation level. The solution is therefore to develop an intelligent data reduction scheme on the detector.

ATLAS is developing a two level data reduction scheme called Hardware Tracking for the Trigger that will make it possible to read out 1 event in 10. However, this requires a massive increase in processing capacity for data reconstruction. Furthermore, reconstruction will need to be done within a few microseconds allowing the majority of the data to be buffered in the ITK (the maximum buffering capacity in the readout ASICs of ITK is a few microseconds).

The only way to achieve high throughput, low latency computing is by hardware accelerated processing. We are currently working on hardware acceleration using Field Programmable Gate Arrays (FPGA).

We are investigating with the WADAPT collaboration the use of high frequency wireless transmission to increase the readout bandwidth from the tracker without increasing material from services. Short range links can also open for interlayer communication in the tracker which may be a powerful method to reduce data volume even further using topological intelligence. The goal is to be able in the future to read out all physics events in the tracker. This would be attractive in particular for Dark Matter searches and new exotic physics.

Track triggers for long-lived particles (LPP)

Contact: Rebeca Gonzalez Suarez, Richard Brenner

Most of new physics searches at the LHC target potential new particles that decay very quickly after they are produced. This fact has driven the design of the detectors and the filters set in place to select potentially interesting events for further processing.

However, there are many plausible new physics models that include metastable or stable new particles, with long decay lengths, that is “long-lived” particles (LLP). In particular, LLP appear as Dark Matter candidates or as new particles linked to a Dark Sector potentially accessible at the LHC. Due to the current event selection and reconstruction, we may be missing those signals.
Tracking information could be used to select events with displaced tracks or vertices, or to tag events with “disappearing tracks”.

In the ATLAS group of Uppsala University we are tying together our interest in Dark Matter searches and our expertise in track triggers to design future mechanisms to identify LLPs already at trigger level in the HL-LHC.

Infrastructure and support

Lots of work is needed to keep the ATLAS detector and the surrounding infrastructure going. The following projects are highly important support functions for the research.

Computing GRID

Contact: Mattias Ellert

The development of a distributed computing GRID has been essential for the success of the LHC program. The development has been the path finder for distributed computing now available in commercial clouds and is still the flagship for data processing in sciences. The GRID engine developed by the Nordic countries called ARC is widely recognized and used in computing infrastructures around the world. ARC is undergoing continuous development. We are currently working on improving the support for submitting computing jobs requiring mixed CPU-GPU architectures.

Fake taus

Contact: Arnaud Ferrari

Many searches for new physics and measurements of Standard Model processes in ATLAS have signatures with hadronically decaying tau-leptons, which are reconstructed as a narrow jet with one or three associated tracks. Multi-variate techniques are employed to distinguish such tau objects from quark- or gluon-initiated jets, however there is always a sizeable background for misidentified taus in ATLAS, and simulation is usually not reliable to estimate it. A number of data-driven background estimation techniques have been employed instead. The main challenge of such methods is that quarks and gluons have different probabilities of faking taus, hence it is crucial to estimate the fractions of quark- and gluon-initiated jets in the data sample used in any given analysis prior to computing the fake-tau background. Our group in Uppsala plays a leading role in the ATLAS Fake Tau Task Force (FTTF), created in 2018 to provide central recommendations on data-driven estimation of fake taus, and on their associated systematic uncertainties.

Last modified: 2022-05-09