Image Credit: LA Times |
Deep learning algorithm does as well as dermatologists
in identifying skin cancer
It's
scary enough making a doctor's appointment to see if a strange mole could be
cancerous. Imagine, then, that you were in that situation while also living far
away from the nearest doctor, unable to take time off work and unsure you had
the money to cover the cost of the visit. In a scenario like this, an option to
receive a diagnosis through your smartphone could be lifesaving.
Universal access to
health care was on the minds of computer scientists at Stanford when they set
out to create an artificially intelligent diagnosis algorithm for skin cancer.
They made a database of nearly 130,000 skin disease images and trained their
algorithm to visually diagnose potential cancer. From the very first test, it
performed with inspiring accuracy.
"We realized it was
feasible, not just to do something well, but as well as a human
dermatologist," said Sebastian Thrun, an adjunct professor in the Stanford
Artificial Intelligence Laboratory. "That's when our thinking changed.
That's when we said, 'Look, this is not just a class project for students, this
is an opportunity to do something great for humanity.'"
The final product, the
subject of a paper in the Jan. 25 issue of Nature, was tested against 21
board-certified dermatologists. In its diagnoses of skin lesions, which
represented the most common and deadliest skin cancers, the algorithm matched
the performance of dermatologists.
Why
skin cancer
Every year there are
about 5.4 million new cases of skin cancer in the United States, and while the
five-year survival rate for melanoma detected in its earliest states is around
97 percent, that drops to approximately 14 percent if it's detected in its latest
stages. Early detection could likely have an enormous impact on skin cancer
outcomes.
Diagnosing skin cancer
begins with a visual examination. A dermatologist usually looks at the
suspicious lesion with the naked eye and with the aid of a dermatoscope, which
is a handheld microscope that provides low-level magnification of the skin. If
these methods are inconclusive or lead the dermatologist to believe the lesion
is cancerous, a biopsy is the next step.
Bringing this algorithm
into the examination process follows a trend in computing that combines visual
processing with deep learning, a type of artificial intelligence modeled after
neural networks in the brain. Deep learning has a decades-long history in
computer science but it only recently has been applied to visual processing
tasks, with great success. The essence of machine learning, including deep
learning, is that a computer is trained to figure out a problem rather than
having the answers programmed into it.
We made a very powerful
machine learning algorithm that learns from data," said Andre Esteva,
co-lead author of the paper and a graduate student in the Thrun lab.
"Instead of writing into computer code exactly what to look for, you let
the algorithm figure it out."
The algorithm was fed each
image as raw pixels with an associated disease label. Compared to other methods
for training algorithms, this one requires very little processing or sorting of
the images prior to classification, allowing the algorithm to work off a wider
variety of data.
From
cats and dogs to melanomas and carcinomas
Rather than building an
algorithm from scratch, the researchers began with an algorithm developed by
Google that was already trained to identify 1.28 million images from 1,000
object categories. While it was primed to be able to differentiate cats from
dogs, the researchers needed it to know a malignant carcinoma from a benign
seborrheic keratosis.
"There's no huge
dataset of skin cancer that we can just train our algorithms on, so we had to
make our own," said Brett Kuprel, co-lead author of the paper and a
graduate student in the Thrun lab. "We gathered images from the internet
and worked with the medical school to create a nice taxonomy out of data that
was very messy -- the labels alone were in several languages, including German,
Arabic and Latin."
After going through the
necessary translations, the researchers collaborated with dermatologists at
Stanford Medicine, as well as Helen M. Blau, professor of microbiology and
immunology at Stanford and co-author of the paper. Together, this
interdisciplinary team worked to classify the hodgepodge of internet images.
Many of these, unlike those taken by medical professionals, were varied in
terms of angle, zoom and lighting. In the end, they amassed about 130,000
images of skin lesions representing over 2,000 different diseases.
During testing, the
researchers used only high-quality, biopsy-confirmed images provided by the
University of Edinburgh and the International Skin Imaging Collaboration
Project that represented the most common and deadliest skin cancers --
malignant carcinomas and malignant melanomas. The 21 dermatologists were asked
whether, based on each image, they would proceed with biopsy or treatment, or
reassure the patient. The researchers evaluated success by how well the
dermatologists were able to correctly diagnose both cancerous and non-cancerous
lesions in over 370 images.
The algorithm's
performance was measured through the creation of a sensitivity-specificity
curve, where sensitivity represented its ability to correctly identify
malignant lesions and specificity represented its ability to correctly identify
benign lesions. It was assessed through three key diagnostic tasks:
keratinocyte carcinoma classification, melanoma classification, and melanoma
classification when viewed using dermoscopy. In all three tasks, the algorithm
matched the performance of the dermatologists with the area under the
sensitivity-specificity curve amounting to at least 91 percent of the total
area of the graph.
An added advantage of the
algorithm is that, unlike a person, the algorithm can be made more or less
sensitive, allowing the researchers to tune its response depending on what they
want it to assess. This ability to alter the sensitivity hints at the depth and
complexity of this algorithm. The underlying architecture of seemingly
irrelevant photos -- including cats and dogs -- helps it better evaluate the
skin lesion images.
Health
care by smartphone
Although this algorithm
currently exists on a computer, the team would like to make it smartphone
compatible in the near future, bringing reliable skin cancer diagnoses to our
fingertips.
"My main eureka
moment was when I realized just how ubiquitous smartphones will be," said
Esteva. "Everyone will have a supercomputer in their pockets with a number
of sensors in it, including a camera. What if we could use it to visually
screen for skin cancer? Or other ailments?"
The team believes it will
be relatively easy to transition the algorithm to mobile devices but there
still needs to be further testing in a real-world clinical setting.
"Advances in
computer-aided classification of benign versus malignant skin lesions could
greatly assist dermatologists in improved diagnosis for challenging lesions and
provide better management options for patients," said Susan Swetter,
professor of dermatology and director of the Pigmented Lesion and Melanoma
Program at the Stanford Cancer Institute, and co-author of the paper.
"However, rigorous prospective validation of the algorithm is necessary
before it can be implemented in clinical practice, by practitioners and
patients alike."
Even in light of the
challenges ahead, the researchers are hopeful that deep learning could someday
contribute to visual diagnosis in many medical fields.
Source: Science Daily
Contingent upon the wellspring of information, information bazaars can be sorted as free or ward. artificial intelligence course
ReplyDeleteWow i can say that this is another great article as expected of this blog.Bookmarked this site.. Dr.jill
ReplyDeleteSocial media sites such as Facebook, Twitter, LinkedIn, Yahoo, AOL, Adobe, Accenture, Datagraph, Google, Hotels and Accommodation, Web Alliance, and so are the major known companies dependent completely on Hadoop. machine learning courses in hyderabad
ReplyDeleteIn this way, on the off chance that we needed to have around 100 billion associations with semiconductors, we will require something like 33.333 billion semiconductors.cyber security course in hyderabad
ReplyDeleteIt is an informative post.
ReplyDeleteThanks for the blog post buddy! Keep them coming... Continue reading
ReplyDeleteExamination should be performed in both the upright and supine positions. One of the best predictors of examination accuracy is the length of time spent by the examiner. Breast Cancer Treatment
ReplyDeleteThis is very significant, and yet necessary towards just click this unique backlink: CBD Bath Bombs
ReplyDelete
ReplyDeleteI am impressed by the information that you have on this blog!
"cyber security course in hyderabad
Such an amazing and informative post you shared with us. I am looking for cipd assignment writing help in dubai so I came across this post. I appreciate your writing.
ReplyDelete