The use of machine learning in the medical field is a topic of great interest these days. Radiology is no exception. In fact, neural networks are increasingly used to help radiologists interpret images and make diagnoses. For example, computer vision models can automatically detect and classify different types of tumors on CT scans.
An AI model is only as good as the data it is trained on. In fact, any bias in the data could be amplified by artificial intelligence. Additionally, the majority of data labeling tools do not support medical imaging or volumetric data formats.
That’s why it is so important to process and label the input data the right way.
Let’s say that we want to train an AI model that tells apart COVID cases and non-COVID viral pneumonia based on X-ray images.
Here is an example of what our ideal data labeling workflow might look like:
Step 1: Uploading files to our training data management platform
Step 2: Creating class labels and annotating the relevant areas of our X-rays
Step 3: Going through a review process to assess the quality of the labels
Step 4: Training a new AI model
Step 5: Using the model for automatic detection of different classes (such as types of infection)
It seems to work just fine, doesn’t it?
The instance segmentation model above automatically chooses which area of the X-ray is relevant. It highlights our regions of interest and also assigns the suspected class based on our training data (Covid for COVID infection and Pneumonia for non-COVID viral pneumonia). AI models can detect patterns with mathematical precision and evaluate how accurate (the percentage value) the predictions are.
However, let’s take a step back.
Before we even start to label our data, we should consider several factors. For example, what if some of our patients were connected to electrocardiogram sensors? Or, what if most non-COVID viral pneumonia cases in our sample were among adolescents while the COVID cases were almost exclusively adults? Sometimes, AI models are too closely tuned to your training data, and they don't generalize well to new data.
It is time to take a closer look at some common challenges of radiology data annotation. And how V7 can help you out.
Some challenges are strictly related to limitations of different imaging technologies used in radiology. Segmenting organs and structures in CT, MRI scans, or X-ray images is a difficult task in itself due to the low contrast between different tissue types. For example, it is hard to determine the ground truth for certain types of images, such as those that show a tumor.
On top of that, you have to consider aspects such as
All of these things increase the difficulty of creating your own ML labeling framework from scratch.
Importing, managing, and annotating radiology data is a complex process. And you cannot train a good model without quality assurance and consistency. AI training data needs to be verified, double and triple checked, and it needs to go through several review stages.
Thankfully, there is no need to reinvent the wheel on your own.
V7 can address all of the issues mentioned above.
Additionally, our network of professional labellers for hire includes scientists and radiologists. If you want to get them involved, they can help you review the annotations or handle them from start to finish. Obviously, it is up to you. You can take care of managing your projects completely on your own and within your team.
To label an MRI or CT scan you need much more than a good understanding of the anatomy. A radiologist has to identify various structures and then label them across different dimensions.
Now, if you want to train a machine learning model, you need lots of data. Quite often, this means repeating the annotation process for the hundredth and thousandth time and going through large datasets of medical images. It can be a daunting task.
There are however some strategies and tools that can make the task easier.
Here are some good practices and interesting features that you can use to improve your radiology annotations:
You can add tags or automatically label your data with custom models. For example, you can set up a custom AI object detection model that will automatically recognize images containing artificial cardiac pacemakers or other medical devices.
For this kind of automatic detection, all you need to do is label 100 instances of an object with bounding boxes. Once you do that, train your model in the Models section of the panel (it takes about 15 minutes), and you are good to go. Start the model, add a model stage in your workflow, and you can detect objects in thousands of new images with a click of a button.
These types of auto-annotations are great for mass operations on your datasets. You can use these auxiliary annotations to filter out outliers or to pick a specific category of images.
Some anatomical structures are quite prominent and easy to label. Sometimes, a tumor segmentation task may be very straightforward. But in other cases, labeling lesions and structures seems next to impossible. To improve their visibility you can use several in-built image manipulation tools, filters, and presets.
For example, an ultrasound scan may be easier to interpret and annotate if you change the color scale. With a dental X-ray you may find that the Bone filter makes your images easier to annotate. You can improve the contrast of your images, alter the display thresholds. And the Invert option is a great tool to use if you need to delineate and label subtle differences in brightness. Some readers are more used to working with an inverted PT image.
Consensus stages added to your review process allow for adjudication of your annotations between different readers. You can specify to what degree your data labels need to overlap to be considered valid.
For example, a workflow manager can assign the same annotation task to a junior radiology technician and one of diagnostic radiographers. If there is a high degree of agreement between the annotations, the annotations can be pushed further down the ML training pipeline. On the other hand, if the overlap is low, you can have the annotations reviewed by a senior radiologist. They can adjust the annotations personally or leave their feedback.
The consensus stage is also very useful for testing your new deep learning models. You can check the performance of your model by having it annotate a small subset of data. For example, your model can outline and label a lesion in a series of MRI scans. You will then be able compare the AI-based medical image segmentation with the annotations of a skilled human radiologist.
V7 offers an experience similar to that of your favorite DICOM viewer. If you are working with volumetric data, take advantage of the image registration views in the main annotation panel.
V7 supports customizable layouts, and an MPR hanging protocol for optimal display of volumetric data. Your imaging series will automatically split into axial, coronal, and sagittal views. This way, you can visualize how your labels are applied in 3D space and make sure that they are consistent across different slices. This feature is essential for medical image annotation.
You can change your position in 3D with your mouse or touchpad. There are also additional crosshair overlays, measurement, and navigation tools that will help you navigate different views with ease and know exactly where you are.
Volumetric data (such as cross-sectional images of a CT scan) works like a sequence of images. This means that you can use the same tools that you would normally use for videos and video frames. One of the most important video annotation features this unlocks is interpolation.
Interpolation is a process of creating new annotations that are based on the existing ones. For example, if you have annotated the boundaries of a tumor in two distant CT slices, you can use interpolation to create in-between labels automatically.
In the example above, we started with one annotation of the spinal cord. Then, we adjusted it every couple of slices. As you can see, the annotation morphs smoothly between all slices. All of the additional labels between our key frames were auto-generated.
Many machine learning projects fail because of poor training data management. When you are working with a large dataset, it is important to keep your data well organized. You should have a separate folder or tag for each modality, body part, and disease. This will help you to find the data you need quickly and will make your labeling and training processes more efficient.
V7 also supports DICOM and NIfTI files. You can easily import and export your data in these formats. Additionally, you can use the platform’s built-in tools for image processing and data augmentation.
Proof of concept testing can help you identify any potential problems that need to be addressed before moving forward with your idea. This is especially true when it comes to machine learning and computer vision projects. With so many factors, and so many ways to fail, it’s important to test your ideas before going all in.
Training several small-scale models with different types of training data can help you assess the feasibility of your project. Getting insights from these experiments can help you avoid wasting time and resources. This can mean saving hundreds of annotation hours of expert radiologists and medical professionals.
In V7, training new models from scratch takes only several minutes. And you can start with as little as one hundred annotations for segmentation, object detection, or classification. Overall, V7 can help you move through the proof of concept stage quickly and efficiently. By doing this, you can focus on developing high-quality machine learning models that meet your needs.
If you are interested in learning more about how V7 can help you with your radiology labeling tasks or you want to see the platform in action, book a demo with our team.