Leveraging AI & ML in MedTech: Purpose-Built Applications
A Venture Capitalist recently joked that to fund a startup, all one must do is choose a URL that ends in ‘.ai’. Although he was not serious, it was an acknowledgment that companies pursuing AI are getting much attention, and there is a fear of missing out (FOMO) in the investment community if one of these AI startups develops a ‘killer app’ that achieves unicorn status without them.
As someone who has been developing products for 30 years, one of the most frequent questions clients ask is how to leverage AI in their products to keep them relevant in a world where AI seems to be growing in almost every industry. There is also a belief that Google, Apple, Amazon, Meta, and Microsoft have all of the data, so how can others ever compete in AI?
The good news about data is that every company has deep knowledge in their domain, and has key data related to their business that is different than the data held by the five big technology companies listed above. The real question is whether your data is in a form that is accessible to build models that will be useful in your business.
Many AI applications focus on a classification task where the data must be labeled to make it useful. A classic example is that of collecting and labeling images as was done by Fei Fei Li, to create the ImageNet database which includes 100,000 synonyms known as ‘synsets’, with 1,000 images for each synset. This labeled database of images has been instrumental in advancing the task of object recognition in machine-learning applications since ImageNet began its effort in 2009, and it is largely responsible for how well AI algorithms can recognize cats, dogs, and other objects in images.
While the terms Artificial Intelligence (AI) and Machine Learning (ML) are often used interchangeably, ML is the process of building a model that classifies data into sub-groups or using regression to represent data along a continuum for continuous data such as temperature predictions. AI is the term we assign to the resulting output from the machine-learning model.
It has been said that current AI models function more like a purpose-built appliance than a human brain, because each application gains expertise in a very narrow space, and this does not generalize well to other spaces. To continue the analogy, a dishwasher and a washing-machine are both purpose-built appliances, and one would not be pleased with the results if they put dishes into the clothes washer. This is the case for most AI models as well.
As an example, the Amazon Alexa app excels in natural language processing to play music, answer questions, set timers, and even tell jokes. However, Alexa would require additional training to recognize cats or dogs if a camera were connected to it. While these AI applications offer a convenience for users within narrow spaces, none of them approach the level of generalized intelligence that is common for a three-year old child.
In the spirit of building purpose-built appliances, let’s look at an example of a practical device to track activity for a patient who wishes to improve their health and wellness. This activity tracker is a coin-sized sensor-tile device with Bluetooth Low Energy (BLE) plus a multi-axis inertial measurement unit (IMU) sensor located in a user’s pocket. Let’s consider how to develop the software for this product by traditional methods vs. an ML approach.
Under a traditional model, a programmer would create a function to determine the orientation of the activity tracker before developing functions to capture data generated from a variety of users when they are walking, running, jumping, etc. This effort requires a substantial amount of time as a programmer has to inspect the IMU data for each user, recognize how this data translates into labeled motion-sequences, and then tune the algorithm to recognize each data-sequence as an activity.
When the same activity tracker is built using ML, users are given a mobile app and the sensor-tile device to put in their pocket. When the sensor-tile detects motion, it notifies the mobile app to ask the user what activity they are doing. The user’s response “labels” each activity, and as this is repeated for all users, a trove of labeled data is easily collected from a group of sensor-tile devices. At the end of each day, the sensor-data from all ‘tiles’ is fed back into an ML model along with the labeled activities from the mobile apps, and the revised model is downloaded into all users’ sensor-tile units. Each sensor-tile now recognizes more activities as users continue to select their activity to label the data. As the model becomes more mature, the mobile app may suggest the activity being detected and allow the user to confirm or correct the activity as needed. This is essentially crowd-sourcing the labeling of data, and it results in a more robust activity sensor as more users participate in its use.
This is an example of a supervised ML application with a classifier that learns to correlate IMU data patterns with activities like walking/running/skipping, etc. If it were trained on only one user, it may work reliably with that user, but would probably do a poor job of identifying activities with different users. This is a classic case of what is called “overfitting,” where an ML model does not generalize well to other user’s data. For this reason, it is important to collect data from a large-enough cross-section of users, most of which (80%) become the training set for your model. The balance of the data collected (20%) are saved to become the test set that is used in qualification and testing of the ML algorithm to verify its capabilities on previously untested data.
While this is a fairly simple example, there are many more applications for AI and ML. In my next column, we will review challenges faced by a key player in the mobile ECG market when using ML to analyze ECG waveforms.
Original post: https://www.medtechintelligence.com/column/leveraging-ai-ml-in-medtech-purpose-built-applications/
Post publish date: