NCCID project summaries¶
The core of the NCCID initiative is to provide value to the public in response to the COVID-19 crisis. For this reason, NCCID wishes to share information on how the data is being utilised by approved institutions and researcherts, to inform the wider community of patients, staff and interested public.
Below is a list of the projects currently ongoing.
University College London¶
University College London plans to store the NCCID data in a highly secure XNAT repository to enable imaging-based research at UCL Centre for Medical Image Computing (CMIC). In particular, researchers at CMIC will use these data to build artificial intelligence models to automatically detect COVID-19 patients based on their CT or X-ray images, such that, in the future, these images can be screened automatically before doctors read them. This will significantly save time in managing future outbreaks. The research also involves building computational models to analyse the outcomes for those with confirmed COVID-19 diagnosis, predicting best management for individual patients. These predictions may shorten their hospital stay, reduce complications and even save lives. Finally, the project will investigates methods to deploy these developed models to local hospitals quickly and safely.
University of Cambridge¶
It is strongly believed that early detection of COVID-19 and intervention leads to lower Covid-19 mortality because it enables disease treatment via oxygen therapy and control of spread via isolation. The diagnosis of COVID-19 must be confirmed by reverse-transcription polymerase chain reaction (RT-PCR) or gene sequencing for respiratory or blood specimens. However, testing the general population is proving to be very challenging because of various reasons including limitations of sample collection and transportation, kit performance and availability, limitations in capacity, etc.
Chest scans could include x-rays, CT and MRI scans. Chest CT scans are used to examine lung tissues and often used for further investigation after an abnormal chest x-ray. Chest MRI scans provide a detailed picture of the chest wall, heart, and blood vessels. These scans are carried out routinely for a variety of medical reasons, including preparation for surgery, annual follow-ups, accident and emergency, etc. The creation of computer systems that can automatically process these scans to detect and identify signs of Covid-19 can provide added value for the NHS with no significant additional burden on staff, resources, operational costs, etc. The development of these systems require the implementation of cutting-edge image processing and artificial intelligence technologies.
X-ray images and CT scans can also be useful in monitoring the progression of Covid-19 patients as they can reveal if their lungs are filled with sticky mucus that can lead to breathing problems and provide a benchmark for comparisons with previous scans.
This project aims to find the visual signatures of Covid-19, as they appear in chest scans, that can lead to accurate diagnostic and prognosis for use in hospital settings. Automated imaging algorithms, aided by advanced artificial intelligence techniques, can detect some of the abnormal features appearing in these scans, such as ground glass patterned areas, Ground glass, Crazy paving, Vascular dilatation, Traction Bronchiectasis. These features are generally not specific to Covid-19 and could be seen with other infections. Hence, it is important to develop AI techniques to aid the imaging analysis to increase the accuracy of diagnostic.
University of Bradford¶
Coronavirus Disease 2019 (COVID-19) is highly contagious, and severe cases can lead to pneumonia and ultimately death. The diagnosis can be confirmed by laboratory testing; however, the test has low sensitivity which leads to late diagnosis and treatment. Chest X-rays and CT scans provide valuable diagnostic and monitoring information that can complement the laboratory and clinical data. In this project, we propose to develop an open-source artificial intelligence tool that combines chest imaging data and clinical data to support the diagnosis, triaging and prognosis for COVID-19 in the UK. This will make clinical decisions more efficient, accurate, timely, and potentially cheaper, leading to better patient outcomes.
V7 Labs is predicting the severity and likelihood of ICU admission of patients via chest x-rays using a deep image retrieval method. By collaborating with NHSx, the San Matteo hospital of Pavia, and multiple international collaborators, they have assembled over 9,000 comparable patient cases, of which over 2,000 are confirmed to have COVID-19. Among these V7’s model has learned to assess the patient’s severity, and whether the patient is at risk of entering the ICU solely based on a single chest x-ray and their patient profile. The AI model currently predicts whether a patient will enter the ICU with 85% accuracy when a first x-ray is taken, based on a validation set of 200 unseen cases from Italy, the UK, and US of varying ethnicities and ages.
V7’s approach involved the extraction of the lungs from a patient’s x-ray to avoid any potential forms of bias. By analysing only the status of pixels inside the lungs, V7’s AI is not able to deduct (and therefore “cheat”) whether a patient is supine, elderly, or had their x-ray taken with a portable scanner. This approach solely observes the patient’s normalised lungs, alongside any quantitative readings taken at the hospital.
The image retrieval component of the AI also presents the 5 most-similar cases to the patient, enabling a deeper level of explainability than simple regions of interest (ROI) on the image. This allows clinician to observe the patient pathway taken by other highly correlated cases. By solely relying on one x-ray, the AI model is able to predict ICU admission with an 85% accuracy rate. The team is now adding additional data to this baseline score to push results to an even higher level and detect potential edge cases.
Chest CT scans are used in hospitals across the globe to image the severity of COVID-19 lung involvement and guide the appropriate patient management. Artificial intelligence (AI) designed for radiologists can increase the speed of reporting on these scans and support timely patient triaging.
Aidence has set-up an international consortium, ICOVAI, to create an AI solution for COVID-19 on chest CTs. The consortium is a collaboration between clinical centers, hospitals, AI companies, and distribution partners.
ICOVAI’s AI solution will automatically detect COVID-19 on chest CTs and assess the extent of lung tissue affected. Its quantitative analysis can be used to guide hospital management, such as bed capacity on wards, or predicting the need for ICU care.
The consortium aims to reduce the workload and pressure that the medical staff are facing during the pandemic. The software will be particularly useful when test kits are absent or inconclusive, and when radiologists are unavailable or lack specific COVID-19 training.
To train a well-performing model, ICOVAI is using high-quality datasets from diverse CT scanners, hospitals, and countries. The patient data is anonymised and processed in line with the GDPR. The product will comply with the Medical Device Regulation (MDR), 2017/745, to ensure clinical safety and quality.
The AI solution will be made available not-for-profit for the NHS and European hospitals. The project is backed by the EU.
City Data Science Institute¶
The City Data Science Institute are using the NCCID dataset to develop artificial intelligence systems that offer explainable decision making. More specifically, we are investigating the key radiological findings of Covid-19, how this can change over time, and how this differs from other disease findings upon a chest X-ray. Using state-of-the-art generative networks, we aim to learn more about the Covid-19 disease process and facilitate medical decision making.
The chest X-ray is a readily available investigation and is useful in the identification, severity assessment and monitoring of Covid-19. It is more readily available than CT scanning and able to exclude other important conditions that may present. The use of AI to assess chest X-rays can facilitate medical staff and improve patient outcomes.
We have begun using a particular type of artificial intelligence model that is able to learn what an X-ray should appear like if it were to be healthy or more diseased. This will help the research community learn about subtle disease features and can contribute to more accurate and quicker automatic diagnostics. One of our main goals is for our system to offer a counterfactual explanation as to why the artificial intelligence has made certain decisions. This is vital in developing safe and effective AI.
Medical Analytica Ltd¶
To test, evaluate and further finetune a software which can automatically analyse Chest X-ray images to identify absence or presence of features of COVID-19 infection. The software can detect other lung conditions such as pneumonia caused by viral or bacterial infections. The software is being further developed to identify other key conditions such as lesion and enlarged heart. The software utilises a number of mathematical models for image analysis and is capable of offering a high confidence classification with minimum false positives or false negatives.
Ultimate objective is integrating it into the NHS radiology reporting workflow to provide a prompt computer aided prediction alongside the radiologist’s own report. By providing an extra layer of support to the clinical team, patient triage and access to appropriate treatment can be speed up. Saving time, resources and most importantly, improving patient experience.
Other application of the software is to provide preliminary / indicative prediction to the primary care team in remote locations and community hospitals to assist in identifying in the community cases which may need urgent specialist attention in main hospitals.
Universities of Brighton, Oxford, Glasgow, Lincoln and Sheffield¶
Members of this collaboration were instrumental in winning first place in ‘Coronahackathon’ April 2020, for the development of Machine Learning (ML) and Artificial Intelligence (AI) to predict patients with SARS-CoV-2 virus using full blood count results. SARS-CoV-2 positive patients exhibit a characteristic change in different parameters measured in simple and rapid blood tests to a high accuracy, predicting the virus in regular wards (93-94%) and those in the community (80-86%).
Our project will validate these initial results and enable use in current hospital practice to screen patients and identify those needing full diagnosis for SARS-CoV-2. Expertise in chest images, blood science and modelling ML and AI will develop an innovative tool to upscale screening to identify individuals for full rt-PCR testing of the virus potentially up to one week earlier than rt-PCR, which will allow much faster release of the country (and the world) from lockdown, protection against future waves and future pandemics.
Ashford and St Peter’s Hospitals NHS Foundation Trust¶
Chest X-rays are often one of the first tests used to help guide doctors with deciding how likely a patient is to have COVID-19 and also how severe the infection is. We are aiming to see how accurate chest X-rays are for detecting COVID-19 and telling doctors how severe the disease is. We plan to answer this by having doctors specialising in X-rays assess chest X-rays of patients in the national COVID-19 imaging database and then compare this assessment to clinical details, such as if they had COVID-19 and how well they did in hospital. We hope this will give us a better understanding of chest X-ray accuracy in COVID-19 and appearances that are linked to COVID-19 or more severe infection. We also additionally aim to use Neural networks (advanced computational algorithms) on the chest X-rays in the database to see if these can be used to automatically detect COVID-19 in chest X-rays. This research will hopefully one day help with the development of clinical algorithms and technology that can be used to speed up chest X-ray assessment for COVID-19.
Behold.ai is currently operating existing services within the NHS for the triage of chest X-rays based on the presence of abnormalities, as determined by artificial intelligence algorithms. In this study Behold.ai aims to achieve two key goals.
First, we aim to validate that an algorithm developed for the general detection of abnormalities on plain-film chest X-rays can identify novel forms of abnormality (such as COVID-19) despite being developed prior to that pathology’s initial presentation (in this case, prior to February 2020). We propose that this will show that the use of existing AI-based patient triage can assist healthcare systems in the efficient use of resource and radiological capacity during high stress periods such as pandemic flu, even when novel pathologies are present.
Secondly, in recognition of the increased utilisation of AI across healthcare sectors, we aim to establish a set of best practices for rapidly and regularly updating clinically deployed algorithms in order to enable efficient ‘learning’ of the characteristics of novel pathologies. This will be validated by the ability of existing chest X-ray models to learn the characteristics of COVID-19 as demonstrated in the NCCID. We suggest that this work will ensure that the implementation of AI into health systems improves future pandemic preparedness as well as maximising responsiveness to population health.
The Royal Surrey NHS Foundation Trust¶
Chest imaging is increasingly used as an alternative method for screening COVID-19, with high sensitivity compared to laboratory testing methods. AI tools have the potential to enable fast and accurate diagnosis from chest X-rays (CXR). However, several issues of image quality have been identified as a limitation to the diagnostic performance of CXR, for example, in images acquired on portable machines, and where under-exposure occurs due to patient positioning or patient BMI. Little is known about the impact of image-quality upon the accuracy and sensitivity of AI algorithms, and we propose to investigate this, having been significantly involved in the development of the NCCID database, and having successfully led previous evaluations of AI tools, for example for diagnosis in breast screening.
The aim of this study is to evaluate the impact of image quality on the performance of AI models by analysing performance metrics during training and validation using the NCCID dataset. Published, open-source models shown to classify COVID-19 on CXR will be identified and the best-performing algorithm selected by assessing performance on the NCCID test dataset. The NCCID training and test datasets will be analysed using existing image analysis tools developed in-house at RSNFT, enabling the categorisation of the NCCID dataset according to image quality, and the dataset will be stratified into datasets of different image quality. The AI model will be retrained and tested on different combinations of training and test data image-quality and the impact of image-quality upon the accuracy and sensitivity of the model evaluated.