Skip to main content

Comparative evaluation and performance of large language models on expert level critical care questions: a benchmark study

Abstract

Background

Large language models (LLMs) show increasing potential for their use in healthcare for administrative support and clinical decision making. However, reports on their performance in critical care medicine is lacking.

Methods

This study evaluated five LLMs (GPT-4o, GPT-4o-mini, GPT-3.5-turbo, Mistral Large 2407 and Llama 3.1 70B) on 1181 multiple choice questions (MCQs) from the gotheextramile.com database, a comprehensive database of critical care questions at European Diploma in Intensive Care examination level. Their performance was compared to random guessing and 350 human physicians on a 77-MCQ practice test. Metrics included accuracy, consistency, and domain-specific performance. Costs, as a proxy for energy consumption, were also analyzed.

Results

GPT-4o achieved the highest accuracy at 93.3%, followed by Llama 3.1 70B (87.5%), Mistral Large 2407 (87.9%), GPT-4o-mini (83.0%), and GPT-3.5-turbo (72.7%). Random guessing yielded 41.5% (p < 0.001). On the practice test, all models surpassed human physicians, scoring 89.0%, 80.9%, 84.4%, 80.3%, and 66.5%, respectively, compared to 42.7% for random guessing (p < 0.001) and 61.9% for the human physicians. However, in contrast to the other evaluated LLMs (p < 0.001), GPT-3.5-turbo’s performance did not significantly outperform physicians (p = 0.196). Despite high overall consistency, all models gave consistently incorrect answers. The most expensive model was GPT-4o, costing over 25 times more than the least expensive model, GPT-4o-mini.

Conclusions

LLMs exhibit exceptional accuracy and consistency, with four outperforming human physicians on a European-level practice exam. GPT-4o led in performance but raised concerns about energy consumption. Despite their potential in critical care, all models produced consistently incorrect answers, highlighting the need for more thorough and ongoing evaluations to guide responsible implementation in clinical settings.

Introduction

Large Language Models (LLMs), a subclass of artificial intelligence (AI) systems that can generate human-like natural language responses to text, hold great promise in supporting healthcare professionals with administrative tasks, such as summarizing clinical notes and drafting answers to patient questions [1]. Additionally, when aiding healthcare professionals with informed medical decisions, we recognize them as AI—clinical decisions support (AI-CDS) tools [2, 3]. Given the complexity and high stakes of such decision-making, ensuring the accuracy and reliability of these systems is paramount, particularly within domains that require highly specific expertise [4].

Clinical decision-making is a multifaceted process that requires medical knowledge, clinical reasoning, the capacity to integrate and synthesize information from various disciplines, and the ability to apply evidence-based practice. However, as a first step for LLMs towards being effective in clinical decision-making, possessing inherent medical knowledge is a fundamental requirement. In recent studies, the medical performance of non-domain specific LLMs was evaluated and findings showed that they can pass the United States Medical Licensing Examination (USMLE) [5, 6]. However, to move towards utilizing LLMs for clinical decision-making in highly specific healthcare fields, where clinical decisions are strongly dependent on a continuous influx of new clinical trial data, it is critical to benchmark LLM models and describe their performance characteristics to guide their safe application in more specific and conceptually complex medical areas. Recent studies evaluated these models in the fields of nephrology and oncology, benchmarking several LLMs using multiple choice questions (MCQs) in their respective field, in which the GPT-4 model typically outperformed the other models [7,8,9]. Similar studies in other specialties have shown comparable trends, with newer models generally performing superiorly [10, 11].

In the critical care medicine, unique challenges present themselves that require rapid decision-making, comprehensive knowledge, and the ability to integrate complex information from multiple organ systems [12]. Currently, a structured assessment of the performance of LLMs in critical care medicine is lacking.

The aim of this study was therefore to evaluate and compare the performance of various LLMs on answering high level critical care questions and compare their performance to human physicians. We also determined costs as a proxy for energy consumption. Our goal was to understand the accuracy, consistency, and limitations of LLMs applied to critical care medicine to guide future studies and applications.

Methods

We performed this cross-sectional comparative study in October 2024. The study was exempt from research ethics board approval and the need for informed consent in accordance with European law, given the lack of involvement of human participants or patient data. We utilized the TRIPOD-LLM guideline for reporting [13].

Dataset

This study utilized 1181 critical care MCQs from the gotheextramile.com (GETM) dataset [14]. GTEM is a comprehensive dataset, designed to assess critical care knowledge at the European Diploma in Intensive Care (EDIC) examination level. Similar to the EDIC exam, the dataset included two types of questions: 74.5% (n = 880) were Type A MCQs (single best answer from four options) and 25.5% (n = 301) were Type K MCQs (four or five true or false statements). The GTEM dataset covers a wide range of critical care domains, which can be found in Supplement A. Due to the nature of the specific LLMs tested, image-based questions were omitted.

Human comparison

For the comparison with human physicians, we evaluated the LLMs on a practice exam at EDIC examination level, consisting of 77 questions, of which 68.8% (n = 53) were Type A and 31.2% (n = 24) were Type K MCQs. For this exam, performance was calculated using a scoring system, in which false answers would deduct points. Human comparators were GTEM subscribers who took the practice exam in 2023 or 2024.

GTEM’s subscriber base spans roughly 50 countries, with India accounting for the largest portion (40–50%). European nations, notably the Netherlands, Denmark, Switzerland, Germany, and the UK, make up the second-largest group, while the United States has minimal representation. The subscriber base consists of physicians, both residents and consultants, predominantly those preparing for the written EDIC exam. Subscribers typically have backgrounds in anesthesiology, with 3 to 5 years of experience in anesthesiology and 1 to 5 years in intensive care. Approximately 30% are repeat subscribers.

Large language models

In total, five LLMs were evaluated: four proprietary foundation models, i.e. GPT-4o, GPT-4o-mini and GPT-3.5-turbo, developed by OpenAI, and Mistral Large 2407 (Mistral AI), and one open-source model, being Llama 3.1 70B (Meta). They were accessed via the Microsoft Azure OpenAI platform. For each model, the temperature setting, which controls the randomness of the model’s outputs, was set to 0, to achieve the most consistent results.

Prompting methods

The prompts were written by a team of experts in medical prompt engineering in an iterative manner. The final prompt used in this study is available in Supplement B. It was important for the final prompt to have utilized zero-shot prompting, meaning that the models are presented with critical care MCQs without prior fine-tuning on medical datasets or knowledge of example questions. This approach aims to assess the models’ inherent critical care knowledge and provide accurate answers based solely on their pre-trained knowledge.

Evaluation

The performance of the LLMs was assessed using multiple evaluation metrics. The primary outcome measure was the overall accuracy, calculated as the percentage of correctly answered MCQs, which we compared to random guessing (e.g., 25% accuracy for questions with four options).

Secondary outcomes included consistency, domain-specific accuracy, costs as a proxy for computing resources or energy use and performance of the LLMs compared with human physicians on a practice exam.

A consistency check was performed by presenting the models with repeated questions to assess the reliability of model responses. One hundred randomly selected questions from the dataset were given to each model 10 times. A model’s response was considered consistent if it provided the same answer for at least 8 out of 10 repetitions (80% threshold) of a given question. The consistency score for each model was calculated as the percentage of questions for which the model demonstrated consistency, distinguishing between consistently correct and consistently incorrect responses.

Domain-specific performance was evaluated by categorizing questions according to critical care topics and calculating accuracy rates for each domain.

Cost per model were calculated directly using the Microsoft Azure OpenAI platform. Pricing is based on the amount of input and output tokens. Efficiency scores were calculated by dividing the performance of the model with its cost.

Statistical analyses

All statistical analyses were performed with Python version 3.11, utilizing libraries such as Numpy, Pandas, Matplotlib, Seaborn and Statsmodels. Continuous variables were summarized using means and standard deviations (SD) or medians and interquartile ranges (IQR), based on the distribution of the data. Categorical variables were summarized as percentages.

The primary outcome measure, overall accuracy, was calculated as the percentage of correctly answered MCQs for each model. Comparative analyses between the performance of the LLMs against random guessing and human physician comparators were conducted using Z-tests for proportions. This method was selected due to the large sample size and binary nature of outcomes (correct/incorrect), allowing for reliable comparisons against fixed performance benchmarks (i.e. random guessing or human comparators).

A p value of < 0.05 was considered statistically significant.

Results

Our results show that the GPT-4o model was the most accurate, correctly answering 93.3% of the questions. For the other models, GPT-4o-mini answered 83.0%, GPT-3.5-turbo 72.7%, Llama 3.1 70B 87.5%, and Mistral Large 2407 87.9% of the answers correctly (Table 1). Considering the number of questions in the dataset and the choices per question, we calculated a performance of 41.5% by random guessing. When compared to random guessing, all models show statistically significant superior performance (p < 0.001).

Table 1 Performance (percentage answered correctly), cost (as a proxy for energy use) and efficiency scores for the various LLMs on all MCQs. Efficiency is the ratio of performance (accuracy) to cost, reflecting the balance between performance and resource usage

All models demonstrated high consistency in their performance (Table 2). Mistral Large 2407 and GPT-4o showed the highest consistency, scoring 100% and 96.0% respectively, providing identical answers in at least 8 out of 10 repetitions for nearly all 100 questions. GPT-4o-mini and Llama 3.1 70B followed closely, scoring 93.0% and 92.0%, while GPT-3.5-turbo showed the lowest consistency at 74.0%. We assessed if the consistent answers were correct or incorrect, relating accuracy and consistency. A high consistently correct score shows that a model is not only accurate but also reliably produces the correct answer when asked multiple times. GPT-4o showed the highest consistently correct score of 88.5% and GPT-3.5-turbo had the least score of 67.6% (Table 2).

Table 2 Consistency scores for all models. Consistency reflects the percentage (%) of questions for which a model provided the same answer in at least 8 out of 10 repetitions, from a random subset of 100 questions. Consistently correct and consistently incorrect scores indicate the proportion of these responses that were accurate or erroneous, respectively

Furthermore, we evaluated the performance for each model for the various critical care subdomains (Fig. 1). For each subdomain, the GPT-4o model scored the highest. Over the various subdomains, all models showed low variability, with a standard deviation of 2.3 for GPT-4o, 3.7 for GPT-4o-mini, 3.5 for GPT-3.5-turbo, 2.6 for Llama 3.1 70B, and 3.7 for Mistral Large 2407. This indicates consistent subdomain knowledge in all tested models. The lowest scoring domains differed between the various tested models, indicating no specific gap in medical subdomain knowledge across LLMs.

Fig. 1
figure 1

Performance per subdomain

As a proxy for computing resources, i.e. energy use, token cost was calculated per model for all 1181 questions. The GPT-4o-mini model was the cheapest, costing €0.14 to run, whereas the GPT-4o model was the most expensive (€3.60). GPT-4o-mini achieves the strongest balance between energy use and performance, offering close-to-top-tier results while requiring much less computational power, reflected in its low cost (Fig. 2). All costs and efficiency scores can be found in Table 1.

Fig. 2
figure 2

Comparing performance and cost (as a proxy for energy use) of for the different LLMs

In the EDIC-like practice exam comparison with human performance, GPT-4o had the highest performance score of 89.0% and GPT-3.5-turbo had the lowest with 66.5% (Fig. 3). All LLMs demonstrated statistically significant superior performance compared to random guessing, calculated as 42.7% (p < 0.001). Between January 2023 and October 2024, 350 physicians completed the practice exam, with an average score of 61.9%. All LLMs outperformed the human physician comparators. However, in contrast to the other evaluated LLMs (p < 0.001), the performance of GPT-3.5-turbo compared to human physicians did not reach statistical significance (p = 0.196).

Fig. 3
figure 3

Performance score of LLMs versus human physicians on practice exam

Discussion

In this study, we evaluated five LLMs, both proprietary and open source, on a large dataset of critical care questions at the European examination level. Our main findings show that all tested LLMs have a high accuracy and as such an outstanding performance. The GPT-4o model had the best performance with 93.3% of questions answered correctly, and significantly outperforms the human physicians. This indicates that critical care domain knowledge is inherently present in these LLMs, potentially rendering them suitable for prospective clinical trials focused on utilizing LLMs in critical care for clinical decision-making.

When compared to the benchmarking studies performed in other highly specific fields, such as nephrology and oncology [7, 8], our results show significantly higher performance of the different models. It is important to note that the best performing model (GPT-4o) is a different model than the best performing models used in those studies (GPT-4). GPT-4o has been shown to have higher performance than GPT-4, which has also been demonstrated in the medical field. For instance, Liu et al. evaluated the performance of various models, including GPT-4o and GPT-4, on the Japanese Medical Licensing Examination and found that GPT-4o outperformed all other tested models with an accuracy of 89.2%, whereas GPT-4 had a significantly lower performance with 76.8% correct [15].

We also evaluated the consistency of the various models and found very high consistency in at least 2 models, i.e. Mistral Large 2407 and GPT-4o. However, for clinical practice, consistency alone is not sufficient to determine the reliability of a model, as it could be consistently incorrect. Consistently incorrect scores were observed in all models. Notably, despite a consistency of 100%, Mistral Large 2407 was consistently incorrect in 19.0% of cases. The GPT-3.5-turbo model showed both the lowest consistency (74.0%) and was only consistently correct in 67.6% of cases. An LLM used in clinical practice should show both high accuracy and high consistency. However, consistently incorrect scores answers raise critical safety concerns particularly relevant to critical care, where dynamic evidence informs decisions with high-stakes consequences. Future research should evaluate techniques to improve safety and should take consistency into account when performing prospective clinical trials with LLMs.

Additionally, we calculated token cost for all models as a proxy for computing resources or energy consumption and as such sustainability. In LLMs, tokens represent a small piece of text to process the input and generate a response, by predicting one token at a time, sequentially. The computational requirements and, consequently, energy consumption is directly proportional to the number of tokens utilized by a model [16, 17]. Furthermore, the size of the LLM is also a significant contributor to energy consumption, with larger models generally consuming more energy per token than their smaller counterparts [18]. LLM token costs could serve as surrogate for energy consumption and as such an indirect measure of sustainability. The most expensive model was GPT-4o, costing over 25 times more than the least expensive model, GPT-4o-mini. However, GPT-4o-mini was 10.3% less accurate than GPT-4o in answering the MCQs correctly. When balancing costs, and thus energy consumption, with performance, GPT-4o-mini yields the highest efficiency. For the implementation of LLMs in healthcare settings, it is crucial to carefully weigh the trade-offs between energy consumption and performance capabilities, including accuracy and consistency. For use cases where inherent medical domain knowledge might be of less importance, such as using LLMs for administrative support, opting for a more cost-effective model might be a more judicious approach. However, even when not used for clinical decision making, assessing accuracy, consistency and clinical validation of the LLM for each use case remain essential steps for the safe implementation of LLMs in healthcare.

While this the first study in its kind to benchmark various LLMs in critical care against examination-level MCQs, several limitations should be acknowledged. First, our study focused on MCQs and omitted image-based questions, which does not fully capture the complexity and depth of understanding required in real-world clinical decision-making. While MCQs are a useful tool for assessing basic knowledge, we did not evaluate the reasoning process behind its decision-making. As a result, the models’ ability to handle more intricate or open-ended problem-solving scenarios, where clinical reasoning is critical, remains untested. Studies on the clinical reasoning capabilities of LLMs have reported promising results, with the GPT-4 model scoring between 66.7 and 76.0% on various general medical questions and clinical cases [19, 20]. However, they are prone to overconfidence and exhibit cognitive biases [21, 22]. This furthermore underscores the importance of thorough and ongoing clinical validation before use these models as clinical decision support tools. For highly specialized subdomains like critical care, these studies are lacking. However, we are currently conducting an evaluation study on the reasoning capabilities of LLMs in critical care medicine to further explore their potential in clinical decision making.

Second, there is a risk of potential data leakage, meaning that questions from the GTEM dataset may have been inadvertently present in the training set of the models. If any of the test questions or their near equivalents were present in the models’ training data, this could artificially inflate their performance. However, given that GTEM is a protected dataset, this is not considered likely. Last, human comparators were physicians preparing for their exam and may have less critical care knowledge and experience than seasoned critical care specialists. However, participating physicians had 1–5 years of experience in critical care.

Conclusion

Our study showed that all tested LLMs have a high accuracy and consistency in answering critical care medicine questions at European examination level. Additionally, four out of five LLMs demonstrated superior performance compared to human physicians on a practice exam, indicating a substantial potential for utilizing these models in critical care. Notably, the GPT-4o model outperformed all other models in terms of accuracy and consistency, yet it was also the most expensive model, suggesting higher energy consumption. When utilizing LLMs in healthcare environments, it is essential to balance energy usage against performance metrics and selecting a more economical model may be the more prudent choice. Despite their potential in critical care, all models produced consistently incorrect answers, highlighting the need for more thorough and ongoing evaluations, with a focus on clinical reasoning capabilities, to guide responsible implementation in clinical settings.

Availability of data

The questions of gotheextramile.com remain confidential.

Code availability

The prompt is attached in the supplementals. Additional information is available upon request.

References

  1. Schoonbeek RC, Workum JD, Schuit SCE, Doornberg JN, Van Der Laan TP, Bootsma-Robroeks CMHHT. Completeness, correctness and conciseness of physician-written versus large language model generated patient summaries integrated in electronic health records [preprint]. SSRN [Internet]. 2024; Available from: https://ssrn.com/abstract=4835935

  2. Sandmann S, Riepenhausen S, Plagwitz L, Varghese J. Systematic analysis of ChatGPT, google search and llama 2 for clinical decision support tasks. Nat Commun. 2024;15(1):2050.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Chung P, Fong CT, Walters AM, Aghaeepour N, Yetisgen M, O’Reilly-Shah VN. Large language model capabilities in perioperative risk prediction and prognostication. JAMA Surg. 2024;159:928–37.

    Article  PubMed  Google Scholar 

  4. Balta KY, Javidan AP, Walser E, Arntfield R, Prager R. Evaluating the appropriateness, consistency, and readability of ChatGPT in critical care recommendations. J Intensive Care Med. 2025;40(2):184–90.

    Article  PubMed  Google Scholar 

  5. Chen H, Fang Z, Singla Y, Dredze M. Benchmarking large language models on answering and explaining challenging medical questions. ArXiv [Internet]. 2024; Available from: http://arxiv.org/abs/2402.18060

  6. Nori H, King N, McKinney SM, Carignan D, Horvitz E. Capabilities of GPT-4 on medical challenge problems. ArXiv [Internet]. 2023; Available from: http://arxiv.org/abs/2303.13375

  7. Wu S, Koo M, Blum L, Black A, Kao L, Fei Z, et al. Benchmarking open-source large language models, GPT-4 and Claude 2 on multiple-choice questions in nephrology. NEJM AI. 2024;1(2):AIdbp2300092.

    Article  Google Scholar 

  8. Rydzewski NR, Dinakaran D, Zhao SG, Ruppin E, Turkbey B, Citrin DE, Patel KR. Comparative evaluation of LLMs in clinical oncology. NEJM AI. 2024;1(5):aloa2300151.

    Article  Google Scholar 

  9. Longwell JB, Hirsch I, Binder F, Gonzalez Conchas GA, Mau D, Jang R, et al. Performance of large language models on medical oncology examination questions. JAMA Netw Open. 2024;7:e2417641.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Passby L, Jenko N, Wernham A. Performance of ChatGPT on specialty certificate examination in dermatology multiple-choice questions. Clin Exp Dermatol. 2024;49:722–7.

    Article  PubMed  Google Scholar 

  11. Ali S, Shahab O, Al Shabeeb R, Ladak F, Yang JO, Nadkarni G, et al. General purpose large language models match human performance on gastroenterology board exam self-assessments. medRxiv [Internet]. Available from: https://doiorg.publicaciones.saludcastillayleon.es/10.1101/2023.09.21.23295918

  12. James FR, Power N, Laha S. Decision-making in intensive care medicine–a review. J Intensive Care Soc. 2018;19(3):247–58.

    Article  PubMed  Google Scholar 

  13. Gallifant J, Afshar M, Ameen S, Aphinyanaphongs Y, Chen S, Cacciamani G, et al. The TRIPOD-LLM statement: a targeted guideline for reporting large language models use 2 3 affiliations. medRxiv [Internet]. Available from: https://doiorg.publicaciones.saludcastillayleon.es/10.1101/2024.07.24.24310930

  14. Arora S. gotheextramile.com [Internet]. Available from: https://gotheextramile.com/

  15. Liu M, Okuhara T, Dai Z, Huang W, Okada H, Furukawa E, et al. Performance of advanced large language models (GPT-4o, GPT-4, Gemini 1.5 Pro, Claude 3 Opus) on Japanese medical licensing examination: a comparative study. medRxiv [Internet]. 2024;2024.07.09.24310129. Available from: http://medrxiv.org/content/early/2024/07/09/2024.07.09.24310129.abstract

  16. Johannes Husom E, Goknil A, Khin Shar L, Sen S. The price of prompting: profiling energy use in large language models inference. ArXiv [Internet]. 2024 [cited 2024 Nov 20]; Available from: https://github.com/ejhusom/MELODI

  17. Luccioni S, Jernite Y, Strubell E. Power hungry processing: watts driving the cost of AI deployment? 2024 ACM conference on fairness, accountability, and transparency, FAccT 2024. Association for Computing Machinery, Inc; 2024. pp. 85–99.

  18. Samsi S, Zhao D, McDonald J, Li B, Michaleas A, Jones M, et al. From words to watts: benchmarking the energy costs of large language model inference. ArXiv. 2023

  19. Cabral S, Restrepo D, Kanjee Z, Wilson P, Crowe B, Abdulnour RE, et al. Clinical reasoning of a generative artificial intelligence model compared with physicians. JAMA Intern Med. 2024;184:581–3.

    Article  PubMed  Google Scholar 

  20. Savage T, Nayak A, Gallo R, Rangan E, Chen JH. Diagnostic reasoning prompts reveal the potential for large language model interpretability in medicine. NPJ Digit Med. 2024;7(1):20.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Griot M, Hemptinne C, Vanderdonckt J, Yuksel D. Large language models lack essential metacognition for reliable medical reasoning. Nat Commun. 2025;16:642.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  22. Wang J, Redelmeier DA. Cognitive biases and artificial intelligence. NEJM AI. 2024;1(12):Alcs2400639. https://doiorg.publicaciones.saludcastillayleon.es/10.1056/AIcs2400639.

    Article  Google Scholar 

Download references

Funding

None.

Author information

Authors and Affiliations

Authors

Contributions

MvG and DvdS conceived the idea. MvG, DvdS and JW designed the study approach. JW drafted the manuscript. JW and BV performed analyses and have designed the prompt. SA designed the questions. MvG, DvdS, and MG critically reviewed and edited the manuscript. DG supervised the process and reviewed the manuscript. All authors approved and reviewed the final manuscript.

Corresponding author

Correspondence to Michel E. van Genderen.

Ethics declarations

Ethical approval statement

Ethics approval was not required for this study as it did not involve human participants, patient data, or any animal subjects and consisted of review of existing data and literature.

Declaration of generative AI and AI-assisted technologies in the writing process

During the writing process of the manuscript, some authors utilized GPT-4o (OpenAI) for linguistic purposes. All authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Conflict of interest

Sumesh Arora is director of gotheextramile.com and its companion App (Critical Care MCQ) and responsible of hosting and developing the dataset of critical care questions used in this study. The other authors have no conflicts of interest to declare.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Workum, J.D., Volkers, B.W.S., van de Sande, D. et al. Comparative evaluation and performance of large language models on expert level critical care questions: a benchmark study. Crit Care 29, 72 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13054-025-05302-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13054-025-05302-0

Keywords