Home » Uncategorized » ARTIFICIAL INTELLIGENCE IN MEDICAL ROBOTICS – CURRENT APPLICATIONS AND POSSIBILITIES

ARTIFICIAL INTELLIGENCE IN MEDICAL ROBOTICS – CURRENT APPLICATIONS AND POSSIBILITIES

AI applications in healthcare are becoming more common for white-collar automation and diagnostics. However, medical robotics is an area that may be marginally underdeveloped. This is likely because of regulations concerning automated surgery.

In this article, we cover how AI software is finding its way into medical robotics now and how it might in the future with more investment and when the density of AI talent at medical robotics companies increases. Specifically, we explore:

  • AI for Medical Robotics – What’s Possible, and What’s Being Used by healthcare clients right now. We found little to no case studies showing a health network or hospital’s success with AI-based medical robotics.
  • The State of AI at Medical Robotics Vendors, including the AI talent at medical robotics companies and a discussion of how to vet a vendor on whether or not its software is truly leveraging AI.

We begin our exploration of A-based medical robots with an overview of how they’re being used now.

AI for Medical Robotics – What’s Possible and What’s Being Used

Theoretically, multiple approaches to developing AI software could work for automating medical robotics. For example, one could use machine vision to guide the robot to problem areas and make it aware of mistakes or patient bodily reactions.

Currently, the medical robotics sector does not have many visible use cases in terms of fully automated surgery or other medical procedures. This is because regulations dictate that a recognized professional administer these procedures. Issues such as liability are harder to resolve with AI because it is usually unclear exactly how an AI application came to its conclusion.

Most medical robots are used for precision operations during non-invasive surgery. This use case nearly prohibits full automation with AI, as no one likely wants to “let loose” an AI software onto the human body. Additionally, a machine learning model built to operate a medical robot with dozens of moving arms and tools would need to be extensively trained on labeled videos of surgeries. This requires thousands of digitally labeled surgical videos before implementation.

A healthcare company may take months to acquire enough data to properly train a machine learning model to perform robotic surgery well enough that it would not be considered a liability. Even if a company did collect all that data, regulations may still need to change before the software can be used to fully automate surgeries.

That said, there are still medical robots for automating other healthcare processes such as diagnostics. For example, Indian software company Sigtuple purportedly created an AI-based telepathology system that automates their smart microscopes to take pictures and send them to the cloud.

Sigtuple’s software is called Shonit, and it consists of smart microscopes, or microscopes fitted to a movable robotic base which are connected to a smartphone camera. The software runs from an app on the smartphone, which also connects it to the cloud. The microscope slides around on its robotic base, which allows the lens to hover over an area of a sample dish and take multiple pictures.

Those pictures are then saved to the smartphone and sent to the cloud to be labeled. The cloud satellite that receives these pictures uses machine vision to label them according to blood cell count and any anomalies within the blood. Then, the pictures are sent to a remote pathologist who can diagnose based on these pre-labeled high-resolution images. Healthcare company workers using the software would then only need to wait for the pathologist to send a response with their diagnosis.

The 3-minute video below explains how the Shonit software can scan blood smears, send them to the cloud for analysis, and then to a pathologist so they may diagnose any illnesses found:

The software likely uses machine vision to cover all of these processes. First, the smartphone camera with the Shonit app installed would use the presence of blood cells as a good indicator to take a picture. Additionally, a higher concentration of cells or bodily structures at the edge of the lens may prompt the software to move the microscope more in the direction of the sample. The cloud-based portion would of course also use machine vision to count blood cells and recognize anomalies or illnesses.

Although this may seem like a novel use of the different capabilities of machine vision, Sigtuple does not list any results showing success with their software. This is because Shonit is still undergoing a partner-exclusive beta launch. Additionally, Sigtuple does not seem to employ long-standing AI talent at their company.

This lack of evidence is present with most medical robotics companies that purport to use AI, as we will discuss in the next section of this article.

The State of AI at Medical Robotics Vendors

A Google search for the top medical robotics companies that are using AI and ML to automate their robotics solutions provides various company names and articles describing each robot. One may see the leading names in medical robotics, such as Intuitive Surgical and Medrobotics Corporation.

These companies all offer surgical robots that facilitate delicate or non-invasive surgeries. They are made to hold tissue in place and make incisions at the same time using multiple robotic arms. Machine vision software can also be used in robotic camera arms that provide a clearer view of body structures during surgery.

Most surgical robots provide helpful information and recommendations during surgery. This can range from monitoring heart rate and blood loss to recommending where to start cutting to remove a foreign body.

Intuitive’s robot also includes an arm with a camera attached to allow for a closer view of the operation. While Intuitive Surgical and Medrobotics claim to use AI, business leaders may not know the trust indicators that would show it.

These companies exhibit some issues regarding their purported use of AI. Medrobotics does not have a robust host of dedicated AI talent, and this is especially true for PhD level staff. Additionally, each company lacks documentation of a client’s success with any kind of software or robotics solution that shows how the AI software solved a business problem.

Though they do not provide evidence of a healthcare company’s success with the software, their advances in robotic cameras and PhD level staff indicates that they may truly be using AI.

Below is a demonstration video showing Intuitive’s Da Vinci robot stitching a grape back together:

Medrobotics does not show as many trust indicators as Intuitive surgical. Some articles may inadvertently mislead a reader to assume a company such as these use real AI because of carefully selected marketing language such as “augmented intelligence,” or “advanced intelligence.”

The criteria we look for in AI software companies are listed below:

  • Talented AI staff with a significant academic background in Machine Learning, AI, or cognitive science. If there are too few PhDaworking on AI at the company, this is a bad sign.
  • Case studies, customer stories, or detailed press releases that provide evidence of a client company’s success with the software. If the company cannot provide so much as a press release with one statistic about their client’s success, it is not likely the software uses AI or is developed enough to go above and beyond for the customer.
  • A value proposition for the software solution that clearly indicates system requirements and inputs and what the system outputs or provides to the user. If one can identify these from a company website they should also be able to determine if the software is actually based on machine learning.

Information on a company’s staff and possible AI talent within that staff can be found on Linkedin. Talented AI staff will likely have “data scientist,” “AI,” or “machine learning,” in their title and have a PhD in machine learning, cognitive science, or another statistical field. Good signs for AI talent include an AI-specific C-level role in the company and multiple PhD holders across levels of seniority within the AI staff.

In order to find case studies that show a client’s success with the software, one may need to search through a company’s website for extra resources or videos. Some companies do not have any case studies, but still list multiple press releases about their clients’ experiences. Press releases are acceptable in cases where they provide detailed accounts of a client’s use of the software and at least one or two statistics that illustrate success with it.

If a company cannot provide any evidence for the legitimacy of their software, it may be best to direct attention elsewhere even if they have considerable AI talent.

A healthcare robotics company could have multiple reasons for claiming to use AI before they have actually implemented it for any of their solutions. One is that it could help the company find new clients that are eager to implement AI at their companies.

Another reason may be that stretching the truth in this way leads to good press for the company, and this good press and new clientele could lead to acquiring more AI staff who can help build the company’s AI applications to better reflect public perception of the company. Once a company like this has a dedicated AI staff, it is only a matter of time before they begin to test machine learning models for automating medical robots.

A company’s value proposition for their software can also illuminate the nature of how it is made and what it is used for. We focus on the system requirements to run properly and what the software does with those resources to determine if it is likely to be AI. Machine learning-based software requires large amounts of training data which is then used to determine when and how to take the next step in a procedure. If a company never states anything about needing to train the software on a corpus of related data, it is likely that it is a difficult process.

AI developers face challenges in terms of the legality and logistics of installing a robotic surgical assistant. As previously stated, a big challenge for the medical robotics field is the concern surrounding fully automated surgical procedures and the resulting healthcare regulations that may prohibit it.

This challenge will likely be overcome with time as the technology becomes more reliable and the public is more comfortable with allowing a robot to operate on them without human assistance.

Additionally, data scientists and machine learning experts may still be developing the method for training a machine learning model to learn surgical procedures. This could be possible with the right surgical footage labeled according to all present body structures and accurate movement or pulsation of those structures. This would also include visible mechanical structures such as the robot’s arms or a surgical implant.

Labeling a surgical video with that amount of information is surely a challenge, and finding a way to do that efficiently and within a reasonable time frame will likely be how these companies overcome it. Healthcare companies trying to make this a reality may benefit from a series of proofreads and approvals by the experienced AI staff at their business.

We spoke to Yufeng Deng, Chief Scientist of Infervision, a machine vision company for medical diagnostics about how data can be most efficiently collected and used for data science purposed in healthcare. In our interview, Deng spoke about the possibilities of machine vision technology in healthcare and more specifically chest diagnostics.

When asked about how his business goes about gathering data from his relevant staff who do not have time to spend on preparing and labeling data, Deng said:

The quality control [of data] is, I will say the most important piece for a good AI model. So we have this four-step quality control process, where each image is at least labeled by two radiologists respectively and independently, and the third step is we’ll have a more experienced radiologist to look at the previous two annotations, two labelings, and make a final decision if these two labelings don’t agree with each other. On top of that, we have a judge who is usually a more experienced person, and on top of that, we have a fourth step, which is a random check process on every day.

Small errors in creating automated tools such as this could endanger a patient’s life, and no healthcare company would want to risk that from a software vendor. So it follows that working software offerings would be few until these challenges are addressed.

 

Header Image Credit: LoHud