top of page
coverpage.jpg
coverpage01.jpg

Design Process

Group 9.png

1/Discovery

Problem

Design question:

When Microsoft first came to us, we were given a big and super difficult task:

How can we cheaply detect diabetic retinopathy in developing nations like India?

Problem space

What's DR(Diabetic retinopathy)?

ds00447_im02507_r7_retinopathythu_jpg.we
Normal-fundus-LRG.jpg

Normal retina

diabetic-retinopathy-header.jpg

Retina with DR

This is when high blood sugar levels cause damage to blood vessels in the retina. These blood vessels can swell and leak. Or they can close, stopping blood from passing through. Sometimes abnormal new blood vessels grow on the retina.

As the condition progresses, diabetic retinopathy symptoms may include:Spots or dark strings floating in your vision (floaters), Blurred vision, Fluctuating vision, Impaired color vision, Dark or empty areas in your vision, Vision loss.

DR in developing countries

In developing countries the high cost of devices to diagnose diabetic retinopathy is prohibitive, and patients are unlikely to travel those distances to seek out medical help. Diabetes is projected to affect over 366 million people in 2030 according to the WHO [1]. Diabetes can lead to eye diseases such as diabetic retinopathy, cataracts, and glaucoma. Furthermore,

diabetes is the 5th leading cause of all blindness in the world [2]. The global prevalence of diabetic retinopathy is 34.6% affecting over 95 million people [3, 7]. When diagnosed and treated early, the risk of going blind decreases by 95% [8].

problem.png

Redefine design question

After having a brief understanding of the problem, we had a discussion with Microsoft team to redefine the design question to give more eye care access to people in developing nations.  

design question.png

But.........

look at our team, no one has an optical physics background.

team.jpg

So in order to approach the problem properly and come up with design features from scratch, we started a first-round research including secondary research, technology research both for software and hardware, competitive analysis, field research, expert interview, surveys, end user interviews.

2/Define

User need

In this stage, we made decision mostly based on expert’s advice and better user experience.

Competitive analysis

屏幕快照 2020-01-09 上午10.53.22.png

After understanding the problem a little more, we had a meeting with Microsoft. In this meeting, we together brainstormed several possible solutions and options, for example, we can use a phone camera directly or a standalone camera system to generate flashing light. In order to make proper design decision, we investigated both the advantages and disadvantages of all possible solutions.

Technology research

Hardware
-Optics & lenses
-Headset design & facial structures
-Mobile phone & RPi cameras
-IR & RGB imaging

Software
-Improving image quality with GANs
-Computer vision techniques
-Multispectral imaging
-Videography

Possible solutions

1/ Phone camera vs. No phone camera

2/ One eye vs. Two eye

3/ Infrared light vs. White light

4/ Dilation vs. No dilation

5/ Take video vs. Take multiple pictures.  

Field research

field study.png

No dilation

-can't see the fundus

-Very bulky but stable

-Intense white light

​-Need professional photographer

With dilation

-can see the fundus

-Medium size 

-Laser light

-​Need professional photographer

With dilation

-can see the fundus

-small size but very unstable

-Intense white light

-Less professional

Expert interview

api-bioimage-aaron-lee.jpg

Aaron Y. Lee M.D.

Ophthalmology

(eye disorders)

-Attach lens to the phone, it can be very small, but it can only see retina with small field of view like 2-3 degree. Phone camera has an infrared filter filtering out the infrared light.

-Expert level is 240 degrees which is the entire retina. At 130 degree macula centered. 1 degree is 2-3 mm.

-We need to reach 60-70 degrees field of view to reach to 95% of the population screening.

-Fovea is the center of the macula which is an area of the retina. Diabetic retinopathy firsts affects the periphery.

-Most eye diseases we have almost balanced eye issues.

-Two key problems of existing landscape: persons hand and patient is moving. Patient does not know where to look. We can direct gaze with the opposite eye, take images of the other eye.

-We can use infrared light source to capture image instead of the white light.

-User experience is important!

-“Human’s eyes are very sensitive to the intense white light, so we have to dilate patients most of the time.”

Sabesan, Ram.jpg

Ramkumar Sabesan

-We can start with a linear structure to understand basic optical and retina knowledge.

-The biggest challenge of the project is how to reduce glare from the cornea and provide a bigger field of view with safe but cheap material and components.

-Image is 256x256 should be sufficient for the models. 15 degree field by 1536x1536 pixels is minimally sufficient for a doctor. A 5MG raspberry pi camera is sufficient (5 MP- 2560x1920 resolution)

-Polarizers and beamsplitters can help reduce glare.

Ophthalmology

Vision Research Scientist

Eye screening experience survey

Since Dr. Lee mentioned that user experience is very important, we want to know what's current eye screening process and its advantages and disadvantages.

We sent out 2 surveys, one for normal people, the other for diabetic patients. Although we did it through Mturk and some of the results is not convincing, we can still find out some common pain points in the survey.

Likes / good experience

survey2.png

Dislikes / bad experience

survey1.png

End user interview

Dr.Guo

42, he is an experienced ophthalmologist who volunteered to help diagnose eye diseases in Africa for one year.

-People have various eye diopters. Patients with myopia, hyperopia and astigmatism all have different diopters, so doctors need to change the focal length of the equipment.
-Most of the normal people’ fundus photos can be taken in the dark time, but some people can’t, including people with astigmatism and cataract. People with diabetics are more likely to have astigmatism and cataract and the radius of the pupil in the dark is smaller than ordinary people (diabetics will affect the muscle around the eye ), so they need to be dilated.
-It can be troublesome when some of the equipment are broken because there are not sufficient professional maintenance workers
-The Internet there is good enough to send pictures and receive basic information and images.
-Most people are not aware that they have more than one eye diseases, so it’s be hard to diagnose for the first time. And even worse, sometimes they only came to the camp once and never showed up anymore even though they just received a cataract surgery. 

-For giving instructions to untrained volunteers, it is very important to first determine what is the cause for the blurred images, is that because of the equipment, their hands or the patients.

Summary- Stakeholder analysis - Design requirement

painpoints.png

Design feature v1

With these design requirements, we summarized some basic design features.

design feature1.png

However, there were also many constrains, including limited processing power, cheap components, unavoidable intensive white light, limited field of view, not self-help product, less precise and limited adjustment range.

3/Ideation

Experiments

Although we have came up with these design requirements, we still felt blank about how to actually take a fundus image. We gradually found out that the first priority of this project is to have a working hardware or camera. We started with doing experiment using an open-source product "O-Docs".

First Failure

It said that O Docs can take clear fundus images, however.....

IMG_3412.jpg
屏幕快照 2020-01-09 下午8.52.40.png

The problems are:

1. This device is not stable at all, both doctors' hands and the patient's eye are shaking and moving.

2. Tons of glare and cornea reflection because of its linear structure.

3. The image quality was high enough and it's too small.

Preliminary design review

With this prototype zero, we had a preliminary design review with Dr. Ramkumar and Dr.Lee. In that meeting we worked collaboratively out a new T-shape lens system to better reduce glare and improve image quality. In this new design, distance between different components are the most important.

hardware.png

Linear polarizer: only let vertical or horizontal light pass,

filter out one direction.

Beamsplitter:splits a beam of light in two.

Condensing lens: converging beam to illuminate an object

Second failure

In addition, in the following hardworking weeks, we also found that IR (Infrared) light is less painful to the eyes, but the lens system reducing glare generated from IR light is different from that from white light. What's worse, if we only use infrared light, it can't be recognized by Microsoft's machine learning model, which is not ideal.

 

So, we determined to use IR light for alignment(focus on the right spot) and take RGB images with a flashing white light.

Third failure

Although, we have already figured out the ideal T-shape lens systems, but in reality, we still get a lot reflection and we can't focus on the right distance with prototype 2.0.  So we still needed to improve the design and the prototype.

Luckily, we have another mentor Joel Kollin from Microsoft to help us, we had a separate hardware meeting with him every week in Microsoft to work together on getting a high resolution fundus image.

jkollin.jpg

1/Camera needs to be sensitive to the 810 nm(near-infrared) range.
2/1.7 mm lens is too broad, we need a smaller field view. 
3/The current RPi camera can be tried with different M-12 lens to test, such as 6mm or 8mm. 
4/80-20 optical bench can be helpful to test our optical system 

5/IR intensity is too low. Originally 10 mW of intensity coming from 810 nm light but only .1 mW reaches the eye. 

Joel Kollin

Principal Research Hardware Development Engineer at Microsoft 

Design feature v2

But the good news is, the image is way better with the new design, so our had a new design features and touch point analysis to work on software simultaneously. 

design feature2.png

Storyboard

With the new design feature, I draw a storyboard of the main process.

storyboard.png

1/Get training before taking image

2/Do alignment in dark with infrared light

3/Click on "Take image" to activate white flashing

4/Upload to cloud to do classification

5/Send back diagnose results

4/Design

Iterations

Prototype design

Iteration 2

屏幕快照 2020-01-10 下午9.19.25.png
屏幕快照 2020-01-10 下午9.19.31.png
IMG_4902.HEIC

Tested on table to see how clear the image we can get. At the same time, we did some math to find out the proper distance.

Iteration 3

IMG_3499.JPG
IMG_3498.JPG
IMG_3501.JPG

Industrial design

IMG_3505.JPG

UI / picture of the front of the eye

Likes

1/All the components can be detached to small modules. 

2/To make it stable, I made a holder to let it stand alone.

3/I made thread on the camera, it can move back and forth to adjust the focus position of the camera.

4/We also have a rough real-time video frame on raspberry pi.

5/Look at the picture we got, even it is only the front of the eye, but the resolution, glare were way better than before.

Problems

1/ Instead of moving the whole camera, better solution is to change the distance between the lens and the filming sensor. But adjusting the inside lens from outside enclosure is not easy.

2/ We still can't see through the pupil to the fundus, in order to solve the problem, we decided to follow the advice from Joel, to take every part out and place it in an optical bench for more precise adjustment.

First evaluation

With the first working prototype, we scheduled the first evaluation.

Goal: it mainly focused on testing whether users can understand how to use the prototype and guide the patients to see different guide lights to take a good front eye image.

At the same time, we wanted to know what our UI features should we provide to help take images.

Method:

1/In order to let them know what a good image, we gave the participants 4 sample images with different guidelights to re-create and capture.

2/We recorded adjustments, errors, comments, and suggestions.

3/Recruited 8 participants to capture images with our prototype.

屏幕快照 2020-01-11 下午6.04.06.png
屏幕快照 2020-01-11 下午6.04.06.png
IMG_4747.jpg
屏幕快照 2020-01-11 下午6.06.33.png
IMG_4761.HEIC
IMG_4755.HEIC
IMG_4751.jpg
IMG_4748.HEIC

Results:

time.png
error.png

Average time: 5 min

(range of 1m40s-10m)

Average error: 32.5%

(range of 8-41%) 

屏幕快照 2020-01-11 下午9.12.53.png
屏幕快照 2020-01-11 下午9.13.01.png

There is large deviation which indicates different understanding of the prototype  between different people .

We compared the image quality based on the euclidean distance between features of the sample images and their images.So there is a gap between the quality of the images from our baseline sample images.

屏幕快照 2020-01-11 下午8.25.21.png

Evaluation result

bad.png

1/Process of setting up and taking a picture was easy, alignment was hard.

2/Orientation of guide lights are confusing

3/Users want to know or confirm the quality of the image they have taken.

good.png

1/User’s should be taught what is type of picture is acceptable.

2/Guide lights help keep the patient’s eye stable

3/Moving the camera was intuitive

Insights:

website.png

Give real-time instructions and guidelines while taking image

User needs confirmation of whether the image quality is acceptable

adjustment.png

X,Y,Z Rail slide for adjustment

Inner camera lens focus adjustment

software.png

Remove circular glare patterns

Image enhancement using SR-GAN, pix2pix GAN with ground truth from clinical fundus cameras

Iteration 4

IMG_4990.jpg
IMG_4958.jpg
36424701-e1573677442390.jpg
IMG_4877.HEIC

Finally!!! Fundus Image!!! 🎉

After putting all the components in an adjustable optical bench, we changed the distances between components and each component's angle more precisely with the help from Joel. In this professional optical bench, we got our first fundus image although with dilation.🎉🎉🎉

We have figured out the right distance and components for the hardware so far, what left to do is to design the companion app and put all the components into an enclosure. 

Key findings:

     Q: What adjustments are necessary to see the fundus of different patients?
     A: Besides from the camera focal length and the distance from the objective lens to the eye needs to be adjusted. The angle of the beamsplitter also need to be adjustable in a small range for glare reduction.

     Q: What is the acceptable range of measurements to repeatedly gather a clear fundus image?
     A: 38-44mm from the objective lens to the front  of the eye. Camera needs to be focused at 110mm.

In order to provide better user experience as expected, we decided to conduct another round user evaluation.

Second evaluation

With the first working prototype, we scheduled the second evaluation.

Goal: 

It mainly helped us find out whether the optical layout enable non-medical professional to capture a standard quality fundus image and test whether the images can be used for processing.

Method:

1/Based on the findings from the Evaluation 1, "User needs confirmation of whether the image quality is acceptable", we developed a simple imaging instructions software which integrated in form of standard image overlay, quality measure and confirmation.

030463558-768x512.jpg

Results:

time.png

Average time:

6 mins 10s vs 5 mins

屏幕快照 2020-01-11 下午9.12.53.png

Image moves too fast due to instability

error.png

Average error: 20% vs 32.5%

屏幕快照 2020-01-12 上午11.36.12.png

It is helpful with the confirmation and the overlaying sample images

bad.png

1/More guidelines are necessary, for example participants found it helpful to be able to start from back and enter the pupil. A demo exercise will be helpful.

-UI
 

2/Confusion about optical disk and different parts of the eye they are seeing.

--Training

3/ At the same time, the field of view this prototype provide is bigger than before but still less than 60-70 degree.

Overall findings

website.png

Companion app

software.png

Software

adjustment.png

Prototype

1/ Give volunteers some basic training on what's optical disk, what's fundus image, how to focus the camera on the pupil before taking actual images.

2/ Give real-time feedback and guidelines while taking image. Imaging instructions integrated in form of standard image overlay is a good option.

3/ Give confirmation after taking the image.

4/ Switch between infrared led and white light led automatically with script.

5/ In order to satisfy the requirement of 60-70 degree of view, we have to take 2 images, one optical disk centered, one fovea centered and stitch them together with ML algorithm to have a whole fundus image. So, we have to place two guide lights horizontally in a small distance inside the enclosure for patients to stare at and keeping their eyes still.

6/ X,Y,Z Rail slide for stable adjustment.

7/ Able to adjust the focal length of the lens and the angle of the beamsplitter from the outside of the enclosure.

5/Design

Solution

Up to now, we had our final solution, it is a very complicated hardware, software and UI system. I met many difficulties in designing UIs. In order to summarize design features and information architecture of the webapp, I first went through our solution with design features and touch point analysis

Storyboard

storyboard.png

Touch point analysis

touch point analysis.jpg

Information architecture

information architecture.png

4/Design

Iterations

Wireframe

With the information architecture, I designed several wireframes.

wireframe.jpg

UI design v1

With wireframes, I designed first version of the UI with Google material design guidance 

all.png

Training

屏幕录制 2020-01-12 下午5.22.11_2.gif

Image taking

target image.gif

Usability test 1

Goal: 

It mainly tested with non-medical professionals and test their user experience of capturing images.

Method:

1/ Recorded mistakes made
2/ Asked their understanding and opinions on the image taking process.
3/ Give scores on the difficulty of using this app for untrained volunteers

4/ Evaluation plan:  https://docs.google.com/document/d/1IuJtark633ueXjMc9ag_RNGvRBlW-lfuWD9mA7T6lIo/edit?usp=sharing

Results:

time.png

Average time:

Average time to complete the entire process (UI only) was under 6 minutes

error.png

Average error:

On average 0.4 mistakes per person

good.png

1/They all think the process is easy to understand and straightforward. The average difficulty score is 3.7 (1 is the easiest, 10 is the highest)

2/Nice information architecture and the interaction is simple.

3/ Giving quizzes after introducing concept as a method of training is very helpful. The quiz scores were high, all participants scored 100/100.

bad.png

Problems:

1/The image of showing live video and the target image at the same time is not intuitive for users, but they can understand easily how to use it after explanation. In addition, they suggest that we can also show some simpler and easier instructions to help the users do alignment.

2/ They didn't really know which step they currently are, they felt confused when they saw the progress bar.  Our system architecture is very complicated, I need to figure out another way to better tell users where are they now

3/ Two third of the participants want to review the images they took, and they want to know what’s going on in each “processing” page. However, Our goal is improve ease of use for the non-professional users and reduce processing for our system. 

4/ They wanted more confirmation in image taking process.

5/ Average mistakes in all the training sections was 1.5 mistakes per participants. The most common mistakes they made were in the first page where they misunderstand the step to be clickable checkbox. -- UI mistakes

A/B Test 1

In our system, we have 5 big steps and 4 sub steps.

1/ Fill out patient information;

2/ Take first optical disk centered image1;

     2.1 see sample images, 2.2 Do alignment with infrared light, 2.3 Take images with white light, 2.4 Confirmation

3/ Take second fovea centered image2;

     2.1 see sample images, 2.2 Do alignment with infrared light, 2.3 Take images with white light, 2.4 Confirmation

4/ Stitching image1 + image2;

5/ Get diagnose result.

In order to figure out the best way to show the progress, I brainstormed 4 different ways and conducted an A/B test.

屏幕快照 2020-01-12 下午10.06.02.png
Take image2.png

Big step: progree bar, sub steps: dots

overlaying images.png

Big step: step number, sub steps: dots

overlaying images2.png

Big step: Text, sub steps: dots

overlaying images3.png

Big step: step number, sub steps: Text

I asked 8 people's opinions on which one is the most clear one,  they love text more but their preference varied. we finally found out the best one is the integrate one, which shows as follows:

123.png

And figured out the best way is to have

a step number showing big steps on the top and a small indicators showing substeps in the bottom. And at the same time have big text on the page to tell the step.

A/B Test 2

We also did an A/B testing on 3 ways to review the image because there is an inconsistent expectation between users and our goal. Our goal is to improve ease of use for the non-professional users and reduce processing for our system, so I didn't encourage users to review them which costs some processing power.

comfirmation-1.png
comfirmation-v2.png

Show images directly on the left

Show images with a obvious button which is in the same importance level.

comfirmation-v3.png

Show images with a less obvious gray text

In the A/B test, some participants thought the first one shows too much information. Compared to 2, 3 shows a sense of our goal which means I can understand that I am able to review the image if I want but I am not encourage to do so. In addition, they thought that after some practices, as volunteers, they don't really need to review the image every time.

So, in order to achieve our goals and hit the user pain points, we will allow users to review images through a less obvious “See Image” button, option 3.

Major UI design improvement

Based on the finding from first usability test and A/B testing, I made the following improvement accordingly.

01

"Showing live video and the target image at the same time is not intuitive for users, although they can understand easily how to use it after explanation. 

 

They suggested that we can give simpler and easier instructions to help the users do alignment."

Capture Image1.png

Instead of doing the sample images vs live video, we give them two different options to switch on and off:


1/Guide circles to only show the place of optical disk or fovea.
2/Transparent sample images to give a more detailed instructions

step3-image2.png

02

Confusion around the 5 big steps and each of their corresponding 4 substeps when taking actual images.

We did A/B testing on 4 ways to show steps and substeps, and which ones were easiest for the user to understand. 

02improvementbefore.png
02improvement.png

03

Two-thirds of users want to review the images they took.

Our goal is improve ease of use for the non-professional users and reduce processing for our system. 

In order to achieve our goals and hit the user pain points, we will allow users to review images through a less obvious “See Image” button

02improvementbefore.png
03.png

04

Users require more indication of when the image is aligned so they know when to capture the image.

Only when the image is aligned will the “take image” option appear on the screen.

04-1.png
step3-image2.png

05

In the training section, The most common mistakes they made were in the first page where they misunderstand the step to be clickable checkbox.

05-1.png
05-2.png

5/Launch

Final presentation

Final User Interface

bg2.png

LOG in

shadow.png
login2.gif
shadow.png
training2.gif
shadow.png
search2.gif

Training

Search a patient

pre2.jpg

Final Prototype

final prototype.jpg
1 (2 - 1)-3.jpg
IMG_5425.jpg
IMG_5241.JPG
030563851-768x512.jpg
IMG_8503.JPG
bottom of page