Placeholder Image

Subtitles section Play video

  • [MUSIC PLAYING]

  • NATHAN SILBERMAN: Imagine you're in a traumatic accident

  • and end up in the hospital.

  • One of the first things an emergency room doctor may do

  • is take an ultrasound device and use

  • it to inspect your heart, your lungs, your kidney.

  • Ultrasound has become an indispensable window

  • into the human body for clinicians

  • across areas of medicine.

  • As powerful as ultrasound is, access

  • to ultrasound is still limited by a very high price,

  • form factor--

  • it's usually a large cart-based system they have to push

  • around--

  • and then, finally, years of education and training

  • required to use it effectively.

  • As a consequence, 2/3 of the world

  • has no access to medical imaging of any kind,

  • 5,000 children die every day from pediatric pneumonia,

  • and over 800 women die every single day

  • from totally preventable complications relating

  • to maternal health.

  • We need to do better.

  • To address this, Butterfly has developed

  • a handheld, pocket-sized ultrasound

  • device that connects right to your smartphone.

  • At $2,000, a hundredth of the price

  • of a conventional ultrasound system,

  • the Butterfly iQ is a personal ultrasound device,

  • a true visual stethoscope.

  • Butterfly's ambition is to democratize ultrasound.

  • The price and size of the device have solved the cost and form

  • factor problems.

  • But to truly make ultrasound universally accessible,

  • we need to solve two additional problems.

  • The first is guidance--

  • where exactly to place the ultrasound probe.

  • That's typically performed by a stenographer.

  • And then, interpretation-- understanding

  • the content of the image is typically

  • performed by a radiologist.

  • Now, in order to solve the access problem,

  • we can't just scale up education.

  • It doesn't scale fast enough.

  • We need to use machine learning.

  • What you're about to see is our machine learning solution

  • to guidance called, acquisition assistance.

  • After indicating in the app what the user wants to image,

  • the app shows the user a split screen.

  • On the bottom, what you're going to see is the ultrasound image.

  • And on the top is an augmented reality interface

  • that shows the user turn-by-turn visual directions

  • that indicates how exactly to move the ultrasound

  • device in order to acquire a diagnostic image.

  • After acquiring a diagnostic image, we need to interpret it.

  • One interpretation model that we've developed

  • is for ejection fraction, an essential measurement

  • of cardiac health.

  • Ejection fraction captures the ratio of blood volume

  • entering and exiting the left ventricle each time

  • your heart beats.

  • This is currently measured by clinicians by hand

  • tracing the edges of the left ventricle,

  • and then evaluating that change over time.

  • Unfortunately, there's great disagreement,

  • even among expert clinicians, with regard

  • to exactly where to place those tracings or even about which

  • frames that are of sufficient quality

  • for reliable evaluation.

  • This chart shows a handful of frames and the responses

  • we obtained from six different expert stenographers regarding

  • whether a frame has sufficient quality.

  • This kind of disagreement introduces a real problem.

  • How do you train and, more importantly,

  • evaluate a model when you don't have access

  • to a single unambiguous ground truth.

  • In our approach to evaluation, rather

  • than comparing to a single ground truth,

  • we seek statistical indistinguishability.

  • Intuitively, this means that if I

  • were to show you a bunch of different estimates,

  • some from a machine and some from humans,

  • you wouldn't be able to tell the difference.

  • So, for example, on the left, what you're seeing

  • is our model in red, which is distinguishable.

  • It has a very clear upward bias.

  • Whereas on the right, if I were to remove the colors,

  • you wouldn't be able to tell which

  • estimates come from the algorithm and which

  • from the clinician.

  • So how do we actually train a model?

  • We use a conventional encoder-decoder architecture.

  • For each frame, we predict whether the frame quality

  • is sufficiently clear for reliable assessment

  • and also produce a per-pixel segmentation

  • of the cardiac chambers.

  • Now that each frame in the video has been segmented,

  • we select a single heart cycle that is of the highest quality.

  • Finally, we estimate the ejection fraction

  • using the largest and smallest areas produced by our model

  • in that heart cycle.

  • Quantitatively, our model is statistically

  • indistinguishable from human experts.

  • More specifically, it produces estimates

  • that are closer to the average over all the clinicians

  • than any of the individual experts are to that average.

  • Underlying all this infrastructure is TensorFlow.

  • Everything that we do runs in real time on the device.

  • This needs to work whether you're

  • in a subbasement of a hospital or in a remote jungle.

  • We train everything with TensorFlow and compile

  • TensorFlow right into the app using

  • a bunch of custom operations.

  • Finally, we also used TF Serving to improve our labeling

  • and monitoring pipelines.

  • To summarize, Butterfly has developed a handheld ultrasound

  • device that has put a high-end ultrasound

  • cart, a stenographer, and a radiologist into your pocket.

  • This is already being used by expert clinicians.

  • And by solving the access problem,

  • our use of real-time machine learning

  • is making the democratization of ultrasound

  • a reality around the world.

  • Thank you.

  • [APPLAUSE]

  • [MUSIC PLAYING]

  • a

[MUSIC PLAYING]

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it

B2

通過ML實現超聲的民主化(TF Dev Summit '19)。 (Democratizing Ultrasound via ML (TF Dev Summit '19))

  • 2 0
    林宜悉 posted on 2021/01/14
Video vocabulary