Placeholder Image

Subtitles section Play video

  • [MUSIC PLAYING]

  • NATHAN SILBERMAN: Imagine you're in a traumatic accident

  • and end up in the hospital.

  • One of the first things an emergency room doctor may do

  • is take an ultrasound device and use

  • it to inspect your heart, your lungs, your kidney.

  • Ultrasound has become an indispensable window

  • into the human body for clinicians

  • across areas of medicine.

  • As powerful as ultrasound is, access

  • to ultrasound is still limited by a very high price,

  • form factor--

  • it's usually a large cart-based system they have to push

  • around--

  • and then, finally, years of education and training

  • required to use it effectively.

  • As a consequence, 2/3 of the world

  • has no access to medical imaging of any kind,

  • 5,000 children die every day from pediatric pneumonia,

  • and over 800 women die every single day

  • from totally preventable complications relating

  • to maternal health.

  • We need to do better.

  • To address this, Butterfly has developed

  • a handheld, pocket-sized ultrasound

  • device that connects right to your smartphone.

  • At $2,000, a hundredth of the price

  • of a conventional ultrasound system,

  • the Butterfly iQ is a personal ultrasound device,

  • a true visual stethoscope.

  • Butterfly's ambition is to democratize ultrasound.

  • The price and size of the device have solved the cost and form

  • factor problems.

  • But to truly make ultrasound universally accessible,

  • we need to solve two additional problems.

  • The first is guidance--

  • where exactly to place the ultrasound probe.

  • That's typically performed by a stenographer.

  • And then, interpretation-- understanding

  • the content of the image is typically

  • performed by a radiologist.

  • Now, in order to solve the access problem,

  • we can't just scale up education.

  • It doesn't scale fast enough.

  • We need to use machine learning.

  • What you're about to see is our machine learning solution

  • to guidance called, acquisition assistance.

  • After indicating in the app what the user wants to image,

  • the app shows the user a split screen.

  • On the bottom, what you're going to see is the ultrasound image.

  • And on the top is an augmented reality interface

  • that shows the user turn-by-turn visual directions

  • that indicates how exactly to move the ultrasound

  • device in order to acquire a diagnostic image.

  • After acquiring a diagnostic image, we need to interpret it.

  • One interpretation model that we've developed

  • is for ejection fraction, an essential measurement

  • of cardiac health.

  • Ejection fraction captures the ratio of blood volume

  • entering and exiting the left ventricle each time

  • your heart beats.

  • This is currently measured by clinicians by hand

  • tracing the edges of the left ventricle,

  • and then evaluating that change over time.

  • Unfortunately, there's great disagreement,

  • even among expert clinicians, with regard

  • to exactly where to place those tracings or even about which

  • frames that are of sufficient quality

  • for reliable evaluation.

  • This chart shows a handful of frames and the responses

  • we obtained from six different expert stenographers regarding

  • whether a frame has sufficient quality.

  • This kind of disagreement introduces a real problem.

  • How do you train and, more importantly,

  • evaluate a model when you don't have access

  • to a single unambiguous ground truth.

  • In our approach to evaluation, rather

  • than comparing to a single ground truth,

  • we seek statistical indistinguishability.

  • Intuitively, this means that if I

  • were to show you a bunch of different estimates,

  • some from a machine and some from humans,

  • you wouldn't be able to tell the difference.

  • So, for example, on the left, what you're seeing

  • is our model in red, which is distinguishable.

  • It has a very clear upward bias.

  • Whereas on the right, if I were to remove the colors,

  • you wouldn't be able to tell which

  • estimates come from the algorithm and which

  • from the clinician.

  • So how do we actually train a model?

  • We use a conventional encoder-decoder architecture.

  • For each frame, we predict whether the frame quality

  • is sufficiently clear for reliable assessment

  • and also produce a per-pixel segmentation

  • of the cardiac chambers.

  • Now that each frame in the video has been segmented,

  • we select a single heart cycle that is of the highest quality.

  • Finally, we estimate the ejection fraction

  • using the largest and smallest areas produced by our model

  • in that heart cycle.

  • Quantitatively, our model is statistically

  • indistinguishable from human experts.

  • More specifically, it produces estimates

  • that are closer to the average over all the clinicians

  • than any of the individual experts are to that average.

  • Underlying all this infrastructure is TensorFlow.

  • Everything that we do runs in real time on the device.

  • This needs to work whether you're

  • in a subbasement of a hospital or in a remote jungle.

  • We train everything with TensorFlow and compile

  • TensorFlow right into the app using

  • a bunch of custom operations.

  • Finally, we also used TF Serving to improve our labeling

  • and monitoring pipelines.

  • To summarize, Butterfly has developed a handheld ultrasound

  • device that has put a high-end ultrasound

  • cart, a stenographer, and a radiologist into your pocket.

  • This is already being used by expert clinicians.

  • And by solving the access problem,

  • our use of real-time machine learning

  • is making the democratization of ultrasound

  • a reality around the world.

  • Thank you.

  • [APPLAUSE]

  • [MUSIC PLAYING]

  • a

[MUSIC PLAYING]

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it