The same devices used to take selfies and type out tweets are being repurposed and commercialized for quick access to information needed for monitoring a patient’s health. A fingertip pressed against a phone’s camera lens can measure a heart rate. The microphone, kept by the bedside, can screen for sleep apnea. Even the speaker is being tapped, to monitor breathing using sonar technology.
In the best of this new world, the data is conveyed remotely to a medical professional for the convenience and comfort of the patient or, in some cases, to support a clinician without the need for costly hardware.
But using smartphones as diagnostic tools is a work in progress, experts say. Although doctors and their patients have found some real-world success in deploying the phone as a medical device, the overall potential remains unfulfilled and uncertain.
Smartphones come packed with sensors capable of monitoring a patient’s vital signs. They can help assess people for concussions, o remedio ibuprofeno contem dipirona watch for atrial fibrillation, and conduct mental health wellness checks, to name the uses of a few nascent applications.
Companies and researchers eager to find medical applications for smartphone technology are tapping into modern phones’ built-in cameras and light sensors; microphones; accelerometers, which detect body movements; gyroscopes; and even speakers. The apps then use artificial intelligence software to analyze the collected sights and sounds to create an easy connection between patients and physicians. Earning potential and marketability are evidenced by the more than 350,000 digital health products available in app stores, according to a Grand View Research report.
“It’s very hard to put devices into the patient home or in the hospital, but everybody is just walking around with a cellphone that has a network connection,” said Andrew Gostine, CEO of the sensor network company Artisight. Most Americans own a smartphone, including more than 60% of people age 65 and over, an increase from just 13% a decade ago, according the Pew Research Center. The COVID-19 pandemic has also pushed people to become more comfortable with virtual care.
Some of these products have sought FDA clearance to be marketed as a medical device. That way, if patients must pay to use the software, health insurers are more likely to cover at least part of the cost. Other products are designated as exempt from this regulatory process, placed in the same clinical classification as a Band-Aid. But how the agency handles AI and machine learning-based medical devices is still being adjusted to reflect software’s adaptive nature.
Ensuring accuracy and clinical validation is crucial to securing buy-in from healthcare providers. And many tools still need fine-tuning, said Eugene Yang, a professor of medicine at the University of Washington. Currently, Yang is testing contactless measurement of blood pressure, heart rate, and oxygen saturation gleaned remotely via Zoom camera footage of a patient’s face.
Judging these new technologies is difficult because they rely on algorithms built by machine learning and artificial intelligence to collect data, rather than the physical tools typically used in hospitals. So researchers cannot “compare apples to apples” with medical industry standards, Yang said. Failure to build in such assurances undermines the technology’s ultimate goals of easing costs and access because a doctor still must verify results.
“False positives and false negatives lead to more testing and more cost to the healthcare system,” he said.
Big tech companies like Google have heavily invested in researching this kind of technology, catering to clinicians and in-home caregivers, as well as consumers. Currently, in the Google Fit app, users can check their heart rate by placing their finger on the rear-facing camera lens or track their breathing rate using the front-facing camera.
“If you took the sensor out of the phone and out of a clinical device, they are probably the same thing,” said Shwetak Patel, director of health technologies at Google and a professor of electrical and computer engineering at the University of Washington.
Google’s research uses machine learning and computer vision, a field within AI based on information from visual inputs like videos or images. So instead of using a blood pressure cuff, for example, the algorithm can interpret slight visual changes to the body that serve as proxies and biosignals for a patient’s blood pressure, Patel said.
Google is also investigating the effectiveness of the built-in microphone for detecting heartbeats and murmurs and using the camera to preserve eyesight by screening for diabetic eye disease, according to information the company published last year.
The tech giant recently purchased Sound Life Sciences, a Seattle startup with an FDA-cleared sonar technology app. It uses a smart device’s speaker to bounce inaudible pulses off a patient’s body to identify movement and monitor breathing.
Binah.ai, based in Israel, is another company using the smartphone camera to calculate vital signs. Its software looks at the region around the eyes, where the skin is a bit thinner, and analyzes the light reflecting off blood vessels back to the lens. The company is wrapping up a US clinical trial and marketing its wellness app directly to insurers and other health companies, said company spokesperson Mona Popilian-Yona.
The applications even reach into disciplines such as optometry and mental health:
With the microphone, Canary Speech uses the same underlying technology as Amazon’s Alexa to analyze patients’ voices for mental health conditions. The software can integrate with telemedicine appointments and allow clinicians to screen for anxiety and depression using a library of vocal biomarkers and predictive analytics, said Henry O’Connell, the company’s CEO.
Australia-based ResApp Health got FDA clearance last year for its iPhone app that screens for moderate to severe obstructive sleep apnea by listening to breathing and snoring. SleepCheckRx, which will require a prescription, is minimally invasive compared with sleep studies currently used to diagnose sleep apnea. Those can cost thousands of dollars and require an array of tests.
Brightlamp’s Reflex app is a clinical decision support tool for helping manage concussions and vision rehabilitation, among other things. Using an iPad’s or iPhone’s camera, the mobile app measures how a person’s pupils react to changes in light. Through machine learning analysis, the imagery gives practitioners data points for evaluating patients. Brightlamp sells directly to healthcare providers and is being used in more than 230 clinics. Clinicians pay a $400 standard annual fee per account, which is currently not covered by insurance. The Department of Defense has an ongoing clinical trial using Reflex.
In some cases, such as with the Reflex app, the data is processed directly on the phone — rather than in the cloud, Brightlamp CEO Kurtis Sluss said. By processing everything on the device, the app avoids running into privacy issues, as streaming data elsewhere requires patient consent.
But algorithms need to be trained and tested by collecting reams of data, and that is an ongoing process.
Researchers, for example, have found that some computer vision applications, like heart rate or blood pressure monitoring, can be less accurate for darker skin. Studies are underway to find better solutions.
Small algorithm glitches can also produce false alarms and frighten patients enough to keep widespread adoption out of reach. For example, Apple’s new car-crash detection feature, available on both the latest iPhone and Apple Watch, was set off when people were riding roller coasters and automatically dialed 911.
“We’re not there yet,” Yang said. “That’s the bottom line.”
Hannah Norman: [email protected], @hnorms
For more news, follow Medscape on Facebook, Twitter, Instagram, and YouTube.
Source: Read Full Article