国产麻豆精品

Skip to content
Will Your Smartphone Be the Next Doctor鈥檚 Office?

Will Your Smartphone Be the Next Doctor鈥檚 Office?

(Oona Tempest/KHN)

The same devices used to take selfies and type out tweets are being repurposed and commercialized for quick access to information needed for monitoring a patient鈥檚 health. A fingertip pressed against a phone鈥檚 camera lens can measure a heart rate. The microphone, kept by the bedside, can screen for sleep apnea. Even the speaker is being tapped, to monitor breathing using sonar technology.

In the best of this new world, the data is conveyed remotely to a medical professional for the convenience and comfort of the patient or, in some cases, to support a clinician without the need for costly hardware.

But using smartphones as diagnostic tools is a work in progress, experts say. Although doctors and their patients have found some real-world success in deploying the phone as a medical device, the overall potential remains unfulfilled and uncertain.

Smartphones come packed with sensors capable of monitoring a patient鈥檚 vital signs. They can , , and conduct mental health wellness checks, to name the uses of a few nascent applications.

Companies and researchers eager to find medical applications for smartphone technology are tapping into modern phones鈥 built-in cameras and light sensors; microphones; accelerometers, which detect body movements; gyroscopes; and even speakers. The apps then use artificial intelligence software to analyze the collected sights and sounds to create an easy connection between patients and physicians. Earning potential and marketability are evidenced by the more than 350,000 digital health products available in app stores, according to a .

鈥淚t鈥檚 very hard to put devices into the patient home or in the hospital, but everybody is just walking around with a cellphone that has a network connection,鈥 said , CEO of the sensor network company Artisight. Most Americans own a smartphone, including more than 60% of people 65 and over, an increase from just 13% a decade ago, according the . The covid-19 pandemic has also pushed people to become more comfortable with virtual care.

Some of these products have sought to be marketed as a medical device. That way, if patients must pay to use the software, health insurers are more likely to cover at least part of the cost. Other products from this regulatory process, placed in the same clinical classification as a Band-Aid. But how the agency handles AI and machine learning-based medical devices is .

Ensuring accuracy and clinical validation is crucial to securing buy-in from health care providers. And many tools still need fine-tuning, said , a professor of medicine at the University of Washington. Currently, Yang is testing contactless measurement of blood pressure, heart rate, and oxygen saturation gleaned remotely via Zoom camera footage of a patient鈥檚 face.

Judging these new technologies is difficult because they rely on algorithms built by machine learning and artificial intelligence to collect data, rather than the physical tools typically used in hospitals. So researchers cannot 鈥渃ompare apples to apples鈥 with medical industry standards, Yang said. Failure to build in such assurances undermines the technology鈥檚 ultimate goals of easing costs and access because a doctor still must verify results.

鈥淔alse positives and false negatives lead to more testing and more cost to the health care system,鈥 he said.

Big tech companies like Google have heavily invested in researching this kind of technology, catering to clinicians and in-home caregivers, as well as consumers. Currently, in the Google Fit app, users can check their heart rate by placing their finger on the rear-facing camera lens or track their breathing rate using the front-facing camera.

(Hannah Norman/KHN; Eric Harkleroad/KHN; Google Fit app; Adobe Stock)
(Hannah Norman/KHN; Google Fit app; Adobe Stock)

鈥淚f you took the sensor out of the phone and out of a clinical device, they are probably the same thing,鈥 said Shwetak Patel, director of health technologies at Google and a professor of electrical and computer engineering at the University of Washington.

Google鈥檚 research uses machine learning and computer vision, a field within AI based on information from visual inputs like videos or images. So instead of using a blood pressure cuff, for example, the algorithm can interpret slight visual changes to the body that serve as proxies and biosignals for a patient鈥檚 blood pressure, Patel said.

Google is also investigating the effectiveness of the built-in microphone for detecting heartbeats and murmurs and using the camera to preserve eyesight by screening for diabetic eye disease, according to information the company .

The tech giant recently , a Seattle startup with an . It uses a smart device鈥檚 speaker to bounce inaudible pulses off a patient鈥檚 body to identify movement and monitor breathing.

, based in Israel, is another company using the smartphone camera to calculate vital signs. Its software looks at the region around the eyes, where the skin is a bit thinner, and analyzes the light reflecting off blood vessels back to the lens. The company a U.S. clinical trial and marketing its wellness app directly to insurers and other health companies, said company spokesperson Mona Popilian-Yona.

The applications even reach into disciplines such as optometry and mental health:

  • With the microphone, uses the same underlying technology as Amazon鈥檚 Alexa to analyze patients鈥 voices for mental health conditions. The software can integrate with telemedicine appointments and allow clinicians to screen for anxiety and depression using a library of vocal biomarkers and predictive analytics, said Henry O鈥機onnell, the company鈥檚 CEO.
  • Australia-based ResApp Health last year for its iPhone app that screens for moderate to severe obstructive sleep apnea by listening to breathing and snoring. SleepCheckRx, which will require a prescription, is minimally invasive compared with sleep studies currently used to diagnose sleep apnea. Those can cost thousands of dollars and require an array of tests.
  • 叠谤颈驳丑迟濒补尘辫鈥檚 app is a clinical decision support tool for helping manage concussions and vision rehabilitation, among other things. Using an iPad鈥檚 or iPhone鈥檚 camera, the mobile app measures how a person鈥檚 pupils react to changes in light. Through machine learning analysis, the imagery gives practitioners data points for evaluating patients. Brightlamp sells directly to health care providers and is being used in more than 230 clinics. Clinicians pay a $400 standard annual fee per account, which is currently not covered by insurance. The Department of Defense has an ongoing clinical trial using Reflex.

In some cases, such as with the Reflex app, the data is processed directly on the phone 鈥 rather than in the cloud, Brightlamp CEO Kurtis Sluss said. By processing everything on the device, the app avoids running into privacy issues, as streaming data elsewhere requires patient consent.

But algorithms need to be trained and tested by collecting reams of data, and that is an ongoing process.

Researchers, for example, that some computer vision applications, like heart rate or blood pressure monitoring, can be less accurate for darker skin. Studies are underway to find better solutions.

Small algorithm glitches can also produce false alarms and frighten patients enough to keep widespread adoption out of reach. For example, Apple鈥檚 new car-crash detection feature, available on both the latest iPhone and Apple Watch, was set off when people were and automatically dialed 911.

鈥淲e’re not there yet,鈥 Yang said. 鈥淭hat’s the bottom line.鈥