`US 20190216326Al
`
`c19) United States
`c12) Patent Application Publication
`Cross et al.
`
`c10) Pub. No.: US 2019/0216326 Al
`Jul. 18, 2019
`(43) Pub. Date:
`
`(54) MULTISPECTRAL MOBILE TISSUE
`ASSESSMENT
`
`(71) Applicant: MIMOSA DIAGNOSTICS INC.,
`Toronto (CA)
`
`(72)
`
`Inventors: Karen Michelle Cross, Toronto (CA);
`General Leung, Toronto (CA)
`
`(21)
`
`Appl. No.:
`
`16/327,786
`
`(22)
`
`PCT Filed:
`
`Aug. 24, 2017
`
`(86)
`
`PCT No.:
`
`PCT /CA2017 /050998
`
`§ 371 (c)(l),
`(2) Date:
`
`Feb. 22, 2019
`
`Related U.S. Application Data
`
`(60)
`
`Provisional application No. 62/378,939, filed on Aug.
`24, 2016.
`
`Publication Classification
`
`(51)
`
`Int. Cl.
`A61B 5100
`A61B 51145
`A61B 511455
`A61B 90/00
`A61B 90/30
`
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`
`G03B 15103
`G03B 29100
`(52) U.S. Cl.
`CPC
`
`(2006.01)
`(2006.01)
`
`A61B 510075 (2013.01); A61B 51004
`(2013.01); A61B 5114542 (2013.01); A61B
`5114552 (2013.01); A61B 51445 (2013.01);
`G03B 2206/00 (2013.01); A61B 90/361
`(2016.02); A61B 90/30 (2016.02); G03B 15103
`(2013.01); G03B 29100 (2013.01); G03B
`2215/0514 (2013.01); A61B 51447 (2013.01)
`
`(57)
`
`ABSTRACT
`
`A method for performing remote tissue assessment of a
`tissue region includes selecting imaging parameters, obtain(cid:173)
`ing M image datasets of the tissue region when the tissue
`region is illuminated by a light signal having a unique
`discrete wavelength, obtaining a reference image dataset of
`the tissue region without illumination, processing the M+l
`image datasets to obtain M marker maps where each marker
`map corresponds to a different physiological marker, and
`analyzing at least one of the marker maps to determine
`whether a physiological condition exists at the tissue region
`or to monitor the status of an existing physiological condi(cid:173)
`tion at the tissue region. In one or more embodiments, the
`method may be performed by a portable light source unit
`including a housing configured for use with a mobile device
`having a camera, or by a system including a light source
`unit, an optical sensor module, and a mobile device.
`
`Light source unit
`
`10
`
`2~
`
`2
`
`~ - - - - -~
`Light source module
`_,22a
`
`i LED driver 1 ; LED 1 i
`
`28
`
`Interface
`unit
`
`26
`
`20
`
`~--- ,----~'----,
`
`Mobile Device
`32
`
`38
`
`12
`
`Analysis server
`
`14
`
`54
`
`User Interface
`
`56
`
`Operating
`System
`fill
`
`48
`
`Web
`portal
`module
`fil
`
`Image
`processing
`module
`fil
`
`Image
`analysis
`module
`fill.
`
`Petitioner's Exhibit 1006
`Page 1 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 1 of 10
`
`US 2019/0216326 Al
`
`28
`
`Interface
`unit
`
`26
`
`20
`
`Power
`module
`
`controller
`
`Light source unit
`
`10
`
`22
`
`Light source module
`,,,,22a
`
`LED driver 1 ! LED 1 !
`
`2b
`
`LED driver 2
`
`16
`
`24
`Communication
`unit
`
`Mobile Device
`32
`
`camera
`
`User Interface
`
`Analysis server
`
`36
`
`Power
`unit
`
`34
`
`Processing
`Unit
`30
`
`Communication
`unit
`
`48
`
`imaging
`acquisition
`module
`
`40
`storage unit
`
`Operating
`System
`42
`
`Computer
`Programs
`44
`
`Data
`store
`46
`
`FIG. 1
`
`52
`
`✓
`
`Communication
`Unit
`
`54
`, /
`
`I
`
`I User Interface
`I
`I
`5
`6
`I
`50'-.
`storage unit /
`Processing Unit ~
`Operating
`I
`System
`~
`
`Web
`portal
`module
`64
`
`Image
`processing
`module
`66
`
`Image
`analysis
`module
`68
`
`Computer
`Programs
`60
`
`Data store
`62
`
`Petitioner's Exhibit 1006
`Page 2 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 2 of 10
`
`US 2019/0216326 Al
`
`100
`~
`
`150
`~
`
`Select imaging parameters
`
`152
`Light distribution correction V
`
`Obtain image datasets for
`M+1 images for M
`physiological markers
`
`Preprocess obtained
`image datasets to obtain M
`marker maps
`
`•
`•
`t
`
`Analyze marker maps to
`determine the existence or
`status of a physiological
`condition at the tissue region
`
`•
`
`Provide recommendation
`
`102
`
`/
`
`104
`
`/
`
`106
`
`/
`
`108
`
`/
`
`110
`
`/
`
`154
`Non-intensity based image V
`registration
`
`•
`+
`
`156
`V
`
`Image Subtraction
`
`♦
`
`I
`~
`Image denoising
`'-------.....------'v1sa
`
`•
`
`Hyper-spectral unmixing
`
`/160
`
`.__ ________ ____,v 162
`_
`Output denoising
`
`1
`
`FIG. 2
`
`FIG. 3A
`
`Petitioner's Exhibit 1006
`Page 3 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 3 of 10
`
`US 2019/0216326 Al
`
`0.3
`
`202
`
`/
`
`41
`
`2 I
`{IL. .. ,.,., ....... , ... ,.,., ..... :, ...... --.-·.,:· ...... ,::------· .. , ..... , ... --: ..
`
`5{i(?
`
`(>U\)
`
`7{ill
`
`~tl<.i
`
`1i!Ji}
`
`!M>
`7M
`\Va-,,;:kagth foml
`
`9(li.l
`
`+ 0.2
`
`204
`
`/
`
`+ 0.5
`
`i) L.., ... w_,., ... , .•... ..,,.:~~~~==::=
`
`~{){)
`
`!ltlll
`700
`\Vm,,,kn~ih (nm)
`
`FIG.3B
`
`Petitioner's Exhibit 1006
`Page 4 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 4 of 10
`
`US 2019/0216326 Al
`
`C'0
`<O
`N
`
`0
`l!)
`
`N~
`
`0
`
`<D N"'
`
`Ill
`"q" .
`C) -LL
`
`N r--(cid:173)
`N
`
`0
`l!)
`
`N~
`
`(0
`I.!)
`N
`
`<(
`"q" .
`C, -LL
`
`N r--(cid:173)
`N
`
`!t
`C:J :"i
`
`···················~'.<~/
`
`C
`"q" .
`C) -LL
`
`N
`I.!)
`N
`
`0
`l!)
`
`N~
`
`C'0
`<D
`N
`
`~I
`
`0
`(0
`N
`
`~I
`
`/.....
`\
`. I
`,; ;Tr-- w·i\··~·-1
`
`N
`l!)
`N
`
`'· ... ..._ ................................. , ,/: ....................................... :
`~---~-----'' ,,_. -----~---~
`
`i
`
`'<:I"
`~
`
`OJ
`~
`
`I.!)
`N
`
`(0
`N
`
`Petitioner's Exhibit 1006
`Page 5 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 5 of 10
`
`US 2019/0216326 Al
`
`.;:: ca
`lO
`!;!;
`... C)
`:!:; U::
`
`Petitioner's Exhibit 1006
`Page 6 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 6 of 10
`
`US 2019/0216326 Al
`
`. . . . . . . . . . . . . . . . .
`
`- .......... :~·-········· . . . . . . . . . . . . . .
`
`:=;: <r-··.::t:.11·::::. ....... 1s
`
`~t;1~
`
`Jo~
`~/ ~~:
`
`} :i:rr
`:;i:
`:"''·:;
`·:I
`.:'.'.'.;
`t:;·:·
`::·½,..--~ ~!{
`
`C
`Lt) .
`!:2
`LL
`
`Petitioner's Exhibit 1006
`Page 7 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 7 of 10
`
`US 2019/0216326 Al
`
`,....
`LO
`Ci)
`
`<D
`LO
`Ci)
`
`N
`LO
`(")
`
`c.o
`
`Jeif ~
`
`co
`LO
`(")
`
`N
`<D
`(V)
`
`Petitioner's Exhibit 1006
`Page 8 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 8 of 10
`
`US 2019/0216326 Al
`
`379 373 378
`
`384 / 3 7 0
`
`_ _ _ 377
`
`---383
`
`375
`
`381
`
`385
`
`FIG. 7A
`
`371
`
`;
`
`.;:
`
`:I
`~ww., :I
`
`FIG. 78
`
`371
`
`Petitioner's Exhibit 1006
`Page 9 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 9 of 10
`
`US 2019/0216326 Al
`
`386
`
`.••••••••••••••••••••
`
`•
`
`-.-.s.:-:-:«0:-'$:-":-":-":-":❖":-":-":-":-'$
`
`X•-.-.::•-:••:;••••••••••••
`
`373 ----..:il
`374 - - - -@
`375
`
`250
`
`..
`§ttl§ffi:ilffi!:i - - - - 378
`377
`376
`
`FIG. 8
`
`Petitioner's Exhibit 1006
`Page 10 of 26
`
`
`
`Patent Application Publication
`
`Jul. 18, 2019 Sheet 10 of 10
`
`US 2019/0216326 Al
`
`27
`
`Optical
`sensor
`roodu!e
`I
`t
`
`I
`
`(.'.:i::rnmHni<:a!i,~n
`tird
`
`Ugh! :S◊tir{:(3 rw)::fai';~
`l2a
`
`I tfi) <l!°h'l;l{ 'l ! L~ 1 I
`um d,hl<'r 2 j 1..~;1
`" ..
`"
`
`.... ,:·')·~-..
`
`,22N
`
`'~
`-M◊tiiletll<''>'ice
`-~
`)2
`----------·,k"""',,,, ~
`1 ~
`jJB;.
`I
`I
`
`40
`~!~,;~i:i~ un=t 1
`
`Fr◊tl;l%it:g
`tfrl:t
`. .:m
`
`-
`
`OptH .. &ti::ig
`s .... stem
`·42
`
`Con1riitinic .. ~tkm
`,;mi!
`
`Opemting
`St~~tero
`i8
`
`Coi~-1pute.f
`Pit>",¥mm,
`fill
`
`Oat;3 titor~
`lU.
`
`kr:~)Je.
`• pr,,o~:s;:s;,ing
`
`~ll~b
`pott::i!
`n:,o<lul~
`ii
`
`!m:s;g~
`.;ln~:ii:fti~>
`m«M~
`f&
`
`<:-0rr1p:,.rti;ir
`Pn::-0,rnm:s
`.:i4
`
`O.it.a
`::;.!~n~
`ii
`
`FIG. 9
`
`Petitioner's Exhibit 1006
`Page 11 of 26
`
`
`
`US 2019/0216326 Al
`
`Jul. 18, 2019
`
`1
`
`MULTISPECTRAL MOBILE TISSUE
`ASSESSMENT
`
`FIELD
`
`[0001] Various embodiments are described herein for a
`device, method and system allowing for the remote assess(cid:173)
`ment of a tissue region.
`
`BACKGROUND
`
`[0002] Patients with diabetes are 20 times more likely to
`have a lower limb amputation than the general population.
`This amputation rate can be increased by 50% if the patient
`has a Diabetic Foot Ulcer (DFU). Surgical limb salvage is
`only possible if the patient presents before irreversible tissue
`damage. However, there are currently no non-invasive tools
`that can give an early warning about the tissue health of
`diabetic legs and feet. This means that patients come to the
`hospital in many cases too late to save their foot or toes. If
`there were an early warning tool, it might be possible to
`detect the disease earlier, intervene earlier, save more limbs,
`and improve these patient's lives.
`[0003] This is important since diabetic patients are 20
`times more likely to be hospitalized with a non-traumatic
`lower limb amputation. The presence of a diabetic foot ulcer
`(DFU) increases the risk of amputation by 50%. These
`patients typically present with micro-angiopathic disease,
`extensive arterial calcification and concomitant large vessel
`peripheral vascular disease. These features lead to chronic
`ischemia of the lower extremities precluding these patients
`to limb ulceration. In addition, patients with diabetes have
`peripheral neuropathy which means they lack sensation in
`their feet and cannot protect themselves from injury. For
`example, a tight fitting shoe causing a blister or pressure
`from stepping on a rough or sharp object is not felt by the
`patient and leads to small abrasions to the skin which
`eventually can progress to DFU's.
`[0004] Early diagnosis and management of DFUs can
`mean the difference between life and death for diabetic
`patients. The one year mortality rate is 30% for diabetics
`whom require lower extremity amputation. This is more than
`the lifetime risk of dying from cancers of the breast and
`colon. The 5 year mortality rate post amputation in patients
`with diabetes is 70%. Approximately, 85% of patients in the
`hospital with concomitant lower extremity amputations have
`diabetes. There are currently no non-invasive techniques for
`monitoring the tissue health of the lower extremity. Personal
`glucometers revolutionized diabetes care 50 years ago by
`permitting "at home" monitoring of blood glucose and being
`able to have closer glycemic control. A personal diagnostic
`device for "foot health" monitoring may have the same
`impact on the incidence of DFU and DFU-related amputa(cid:173)
`tions through early recognition and intervention.
`
`SUMMARY OF VARIOUS EMBODIMENTS
`
`[0005]
`least one embodiment
`In a broad aspect, at
`described herein provides a portable light source unit for
`illuminating a tissue region for imaging, wherein the light
`source unit includes a housing being configured for use with
`a mobile device having a camera, the housing having a base
`member with a first aperture and at least two opposing side
`walls disposed inwardly and being spaced apart to releasably
`receive edges of the mobile device and the first aperture
`having a location that is aligned with a location of the
`
`camera of the mobile device when the portable light source
`unit is mounted on the mobile device; a light source module
`disposed along a portion of the base member and oriented to
`generate a light signal in a common direction with the
`camera, the light source module being located to be outside
`of the field of view of the camera; and a controller that is
`operatively coupled with the light source module and a
`processor of the mobile device to generate a light control
`signal for controlling the light source module to generate
`light to illuminate the tissue region prior to the camera of the
`mobile device obtaining an image of the tissue region.
`[0006]
`In at least some embodiments, the controller
`includes a communication unit for wirelessly communicat(cid:173)
`ing with the mobile device. The communication unit may
`include a Bluetooth radio.
`[0007]
`In at least some embodiments, the portable light
`source unit further includes a power module having a battery
`and a voltage regulator to provide power to electronic
`components of the portable light source unit.
`[0008] Alternatively, in at least some embodiments, the
`portable light source unit further is coupled to the mobile
`device to receive power therefrom to provide power to
`electronic components of the portable light source unit.
`[0009]
`In at least some embodiments, the light source
`module, the controller and the power supply unit are
`mounted on a circuit board that is disposed within the
`housing.
`[0010]
`In at least some embodiments, the light source
`module is configured to generate the light signal to have one
`ofN single discrete wavelengths in the visible light and near
`infrared spectrum.
`[0011]
`In at least some embodiments, the light source
`module includes N LED drivers and N LEDs, where each
`LED is configured to generate a light signal having a single
`discrete unique wavelength in the range of about 600 nm to
`1,000 nm.
`[0012]
`In at least some embodiments, the light source
`module includes a second aperture that is aligned with the
`camera of the mobile device when the portable light source
`unit is mounted on the mobile device and the LEDs are
`arranged in a circular pattern around the second aperture.
`[0013] The N discrete wavelengths of the light source
`module may generally be selected based on N physiological
`markers of interest that are measured when the tissue region
`is imaged.
`[0014]
`least one embodiment
`In another aspect, at
`described herein provides a method for performing remote
`tissue assessment for a tissue region, wherein the method
`involves: selecting imaging parameters; obtaining M image
`datasets of the tissue region when the tissue region is
`illuminated by a light signal having a unique discrete
`wavelength selected from M discrete unique frequencies;
`obtaining a reference image dataset of the tissue region
`when the tissue region is not illuminated;
`[0015] processing the M+l image datasets to obtain M
`marker maps where each marker map corresponds to a
`different physiological marker; and analyzing at least one of
`the marker maps to determine whether a physiological
`condition exists at the tissue region or to monitor the status
`of an existing physiological condition at the tissue region.
`[0016]
`In at least some embodiments, the method further
`involves providing a recommendation depending on the
`analysis results, the recommendation comprising one of
`
`Petitioner's Exhibit 1006
`Page 12 of 26
`
`
`
`US 2019/0216326 Al
`
`Jul. 18, 2019
`
`2
`
`performing further assessment of the tissue region and
`applying a type of treatment to the tissue region.
`[0017]
`In at least some embodiments, the discrete wave(cid:173)
`lengths are selected in the range of about 600 nm to 1,000
`nm. More particularly, in at least some embodiments, the
`discrete wavelengths include 620, 630, 700, 810, 880 and
`940 nm.
`[0018]
`the analysis
`least some embodiments,
`In at
`involves combining at least two of the marker maps to create
`a combined marker map and performing measurements on
`the combined marker map.
`[0019]
`In at least some embodiments, the physiological
`markers for the marker maps include at least one of total
`hemoglobin, oxygen saturation, methemoglobin, water, and
`melanin.
`[0020]
`In at least some embodiments, the marker map
`includes a water marker map that is used for at least one of
`determining end points of resuscitation, monitoring burn
`edema, monitoring sepsis, and monitoring infection.
`[0021]
`In at least some embodiments, the marker map
`includes a water marker map that is used for at least one of
`skin hydration, skin dehydration, and skin moisture absorp(cid:173)
`tion.
`[0022]
`In at least some embodiments, the marker map
`includes a melanin marker map that is used for at least one
`of skin pigmentation monitoring, skin whitening treatment,
`and treatments for skin hyperpigmentation.
`[0023]
`In at least some embodiments, the marker map
`includes a collagen marker map that is used to monitoring at
`least one of wound healing, scar outcome and scar progres(cid:173)
`s10n.
`[0024]
`In at least some embodiments, the method involves
`generating an alert when the analysis of the at least one
`marker map indicates a sudden change in the physiological
`condition.
`[0025]
`In at least some embodiments, the method involves
`determining a relationship between at least two of the
`marker map; and analyzing the determined relationship to
`determine the composition of blood oxygenation and oxi(cid:173)
`dation for the tissue region.
`[0026]
`In at least some embodiments, the processing
`involves performing multispectral unmixing to generate the
`marker maps.
`[0027]
`In at least some embodiments, the processing
`involves applying at least one of light distribution correc(cid:173)
`tion, non-intensity based image registration, image subtrac(cid:173)
`tion, image denoising and output denoising to improve the
`signal to noise ratio of the generated marker maps.
`[0028]
`In at least some embodiments, the method involves
`using a handheld portable imaging device for imaging the
`tissue region, the handheld portable imaging device includ(cid:173)
`ing a light source unit for generating the light signal to
`illuminate the tissue region and a mobile device having a
`camera for obtaining the image datasets.
`[0029]
`In at least some embodiments, the image datasets
`are sent from the mobile device to an analysis server which
`performs the processing and analysis of the image datasets.
`[0030]
`In at least some embodiments, the mobile device
`performs the processing and analysis of the image datasets.
`[0031]
`In at least some embodiments, the method includes
`using a system for imaging the tissue region, the system
`including a light source unit for generating the light signal
`to illuminate the tissue region, an optical sensor module in
`
`communication with the light source unit for obtaining the
`image datasets, and a mobile device in communication with
`the light source unit.
`[0032]
`least one embodiment
`In another aspect, at
`described herein provides a system for remote tissue assess(cid:173)
`ment of a tissue region, wherein the system includes: a light
`source unit for generating a light signal to illuminate the
`tissue region, the light signal having a unique discrete
`wavelength selected from M discrete unique frequencies;
`and a mobile device having a camera for obtaining M image
`datasets of the tissue region when the tissue region is
`illuminated by the light signal and obtaining a reference
`image dataset of the tissue region when the tissue region is
`not illuminated, wherein the system is configured to: process
`the M+l image datasets to obtain M marker maps where
`each marker map corresponds to a different physiological
`marker; and analyze at least one of the marker maps to
`determine whether a physiological condition exists at the
`tissue region or to monitor the status of an existing physi(cid:173)
`ological condition at the tissue region.
`[0033] The system may further include an analysis server
`portal.
`[0034] According to one or more embodiments, a system
`for remote tissue assessment of a tissue region includes a
`light source unit for generating a light signal to illuminate
`the tissue region, the light signal having a unique discrete
`wavelength selected from M discrete unique frequencies, an
`optical sensor module for obtaining M image datasets of the
`tissue region when the tissue region is illuminated by the
`light signal and obtaining a reference image dataset of the
`tissue region when the tissue region is not illuminated,
`wherein the optical sensor module is in communication with
`the light source unit; and a mobile device in communication
`with the light source unit, wherein the system is configured
`to process the M+l image datasets to obtain M marker maps
`where each marker map corresponds to a different physi(cid:173)
`ological marker, and analyze at least one of the marker maps
`to determine whether a physiological condition exists at the
`tissue region or to monitor the status of an existing physi(cid:173)
`ological condition at the tissue region.
`[0035] Other features and advantages of the present appli(cid:173)
`cation will become apparent from the following detailed
`description taken together with the accompanying drawings.
`It should be understood, however, that the detailed descrip(cid:173)
`tion and the specific examples, while indicating preferred
`embodiments of the application, are given by way of illus(cid:173)
`tration only, since various changes and modifications within
`the spirit and scope of the application will become apparent
`to those skilled in the art from this detailed description.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[0036] For a better understanding of the various embodi(cid:173)
`ments described herein, and to show more clearly how these
`various embodiments may be carried into effect, reference
`will be made, by way of example, to the accompanying
`drawings which show at least one example embodiment, and
`which are now described. The drawings are not intended to
`limit the scope of the teachings described herein.
`[0037] FIG. 1 is a block diagram of an example embodi(cid:173)
`ment of a system including a portable imaging device having
`a light source unit and a mobile device for remote tissue
`assessment.
`[0038] FIG. 2 is a flowchart of an example embodiment of
`a method for remote optical tissue assessment.
`
`Petitioner's Exhibit 1006
`Page 13 of 26
`
`
`
`US 2019/0216326 Al
`
`Jul. 18, 2019
`
`3
`
`[0039] FIG. 3A is a flowchart of an example embodiment
`of an image processing method that can be used with the
`method for remote optical tissue assessment shown in FIG.
`2.
`[0040] FIG. 3B is a graphical example of multispectral
`unmixing.
`[0041] FIG. 4A is a plan view of an inner surface of an
`example embodiment of a housing for the light source unit
`of FIG. 1.
`[0042] FIG. 4B is a perspective view of the inner surface
`of the housing of FIG. 4A.
`[0043] FIG. 4C is a side view of the bottom edge of the
`housing of FIG. 4A.
`[0044] FIG. 4D is a side view of the housing of the FIG.
`4A.
`[0045] FIGS. SA, 5B, SC, and SD show schematics of an
`example embodiment of electronic components which may
`be used with the light source unit of FIG. 1.
`[0046] FIG. 6 shows an image of an example embodiment
`of a light source module that can be used with the light
`source unit of FIG. 1.
`[0047] FIG. 7Ashows a top image ofan example embodi(cid:173)
`ment of a printed circuit board having a light source module
`disposed thereon. FIG. 7B shows a bottom image of the
`example embodiment of the printed circuit board having the
`light source module disposed thereon.
`[0048] FIG. 8 is an image of an example embodiment of
`a housing of a light source unit.
`[0049] FIG. 9 is a block diagram of an example embodi(cid:173)
`ment of a system including a light source unit, an optical
`sensor module, and a mobile device for remote tissue
`assessment.
`[0050] Further aspects and features of the example
`embodiments described herein will appear from the follow(cid:173)
`ing description taken together with the accompanying draw(cid:173)
`ings.
`
`DETAILED DESCRIPTION OF THE
`EMBODIMENTS
`
`[0051] Various embodiments in accordance with the
`teachings herein will be described below to provide an
`example of at least one embodiment of the claimed subject
`matter. No embodiment described herein limits any claimed
`subject matter. The claimed subject matter is not limited to
`devices, systems or methods having all of the features of any
`one of the devices, systems or methods described below or
`to features common to multiple or all of the devices and or
`methods described herein. It is possible that there may be a
`device, system or method described herein that is not an
`embodiment of any claimed subject matter. Any subject
`matter that is described herein that is not claimed in this
`document may be the subject matter of another protective
`instrument, for example, a continuing patent application,
`and the applicants, inventors or owners do not intend to
`abandon, disclaim or dedicate to the public any such subject
`matter by its disclosure in this document.
`[0052]
`It will be appreciated that for simplicity and clarity
`of illustration, where considered appropriate, reference
`numerals may be repeated among the figures to indicate
`corresponding or analogous elements. In addition, numerous
`specific details are set forth in order to provide a thorough
`understanding of the embodiments described herein. How(cid:173)
`ever, it will be understood by those of ordinary skill in the
`art that the embodiments described herein may be practiced
`
`without these specific details. In other instances, well-known
`methods, procedures and components have not been
`described in detail so as not to obscure the embodiments
`described herein. Also, the description is not to be consid(cid:173)
`ered as limiting the scope of the embodiments described
`herein.
`[0053]
`It should also be noted that the terms "coupled" or
`"coupling" as used herein can have several different mean(cid:173)
`ings depending in the context in which these terms are used.
`For example, the terms coupled or coupling can have an
`electrical connotation. For example, as used herein, the
`terms coupled or coupling can indicate that two elements or
`devices can be directly connected to one another or con(cid:173)
`nected to one another through one or more intermediate
`elements or devices via an electrical signal that can be
`transmitted over a physical wire or cable or transmitted
`wirelessly. In other instances the terms coupled or coupling
`can indicate that two elements are directly mechanically
`connected to one another or mechanically connected to one
`another through another element or linkage.
`[0054]
`It should also be noted that, as used herein, the
`wording "and/or" is intended to represent an inclusive-or.
`That is, "X and/or Y" is intended to mean X or Y or both,
`for example. As a further example, "X, Y, and/or Z" is
`intended to mean X or Y or Z or any combination thereof
`[0055]
`It should be noted that terms of degree such as
`"substantially", "about" and "approximately" as used herein
`mean a reasonable amount of deviation of the modified term
`such that the end result is not significantly changed. These
`terms of degree may also be construed as including a
`deviation of the modified term if this deviation does not
`negate the meaning of the term it modifies.
`[0056] Furthermore, the recitation of numerical ranges by
`endpoints herein includes all numbers and fractions sub(cid:173)
`sumed within that range (e.g. 1 to 5 includes 1, 1.5, 2, 2.75,
`3, 3.90, 4, and 5). It is also to be understood that all numbers
`and fractions thereof are presumed to be modified by the
`term "about" which means a variation of up to a certain
`amount of the number to which reference is being made if
`the end result is not significantly changed, such as 5% or
`10%, for example, as the case may be.
`[0057] The example embodiments of the systems and
`methods described in accordance with the teachings herein
`may be implemented as a combination of hardware or
`software. Accordingly, the example embodiments described
`herein may be implemented, at least in part, by using one or
`more computer programs, executing on one or more pro(cid:173)
`grammable devices comprising at least one processing ele(cid:173)
`ment, and a data store (including volatile and non-volatile
`memory and/or storage elements). These devices may also
`have at least one input device (e.g. a receiver) and at least
`one output device (e.g. a transmitter) or in some embodi(cid:173)
`ments an input/output device such as a transceiver, depend(cid:173)
`ing on the nature of the device. The programmable devices
`may include a processor, a controller, or an Application
`Specific Integrated Circuit (ASIC), for example. Program
`code can be applied to input data to perform the functions
`described herein and to generate output data. The output data
`can be supplied to one or more output devices for outputting
`to one or more electronic devices.
`[0058] There may be some elements that are used to
`implement at least part of one of the embodiments described
`herein that may be implemented via software that is written
`in a high-level procedural language or object oriented pro-
`
`Petitioner's Exhibit 1006
`Page 14 of 26
`
`
`
`US 2019/0216326 Al
`
`Jul. 18, 2019
`
`4
`
`gramming and/or scripting language. Accordingly, the pro(cid:173)
`gram code may be written using an event-driven paradigm
`with objects or any other suitable programming language
`and may include modules or classes, as is known to those
`skilled in object oriented programming. Alternatively, or in
`addition thereto, some of these elements implemented via
`software may be written in assembly language, machine
`language or firmware as needed. In either case, the language
`may be a compiled or an interpreted language.
`[0059] At least some of these software programs may be
`stored on a storage media ( e.g. a computer readable medium
`such as, but not limited to, ROM, FLASH memory and the
`like) or a device that is readable by a general or special
`purpose programmable device. The software program code,
`when read by the programmable device, configures the
`programmable device to operate in a new, specific and
`predefined manner in order to perform at least one of the
`methods described in accordance with the teachings herein.
`[0060] Furthermore, at least some of the programs asso(cid:173)
`ciated with the systems and methods of the embodiments
`described herein may be capable of being distributed in a
`computer program product comprising a computer readable
`medium that bears computer usable instructions for one or
`more processors. The medium may be provided in various
`forms, including non-transitory forms such as, but not
`limited to, one or more diskettes, compact disks, tapes,
`chips, and magnetic and electronic storage. In alternative
`embodiments, the medium may be transitory in nature such
`as, but not limited to, wire-line transmissions, satellite
`transmissions,
`interne
`transmissions
`(e.g. downloads),
`media, digital and analog signals, and the like. The computer
`useable instructions may also be in various formats, includ(cid:173)
`ing compiled and non-compiled code.
`In accordance with the teachings herein, at least
`[0061]
`one example embodiment of a portable imaging device that
`incorporates optical components is provided herein. The
`portable imaging device can be used in a variety of ways.
`For example, the portable imaging device can be used in
`remote assessment of a tissue region for a physiological
`condition and therefore provide an early warning for the
`occurrence of the physiological condition or for a worsening
`status of the physiological condition.
`[0062] The portable imaging device is hand-held and can
`be easily manipulated with one hand in order to obtain
`images for any body part. The device is also easy to use.
`Accordingly, the device can be used by lay people to assess
`tissue health at various remote locations with respect to a
`medical institution ( e.g. a hospital, clinic or the like), such
`as a person's home. The portable imaging device will permit
`for rapid assessment of wounded tissue, non-wounded tissue
`and tissue health, which has been previously unavailable to
`clinicians and patients.
`In an example embodiment, the portable imaging
`[0063]
`device may be incorporated with wound management soft(cid:173)
`ware that implements a wound tracking system that can be
`used to automate follow up appointments and provide early
`warning for the detection of negative tissue changes thereby
`alerting physicians, clinicians or healthcare workers to
`patients requiring further assessment and intervention such
`as dressing, surgery and the like. This will reduce the
`number of hospital visits and permit early detection of tissue
`ischemia.
`In at least one example embodiment described in
`[0064]
`accordance with the teachings herein, image data can be
`
`obtained for a sequence of images of a tissue region for a
`portion of a body part. The sequence of images may be
`obtained while illuminating a tissue region with a sequence
`of different wavelengths of visible and Near InfraRed (NIR)
`light so that each image is obtained while illuminating the
`tissue region with light having a different wavelength. A
`reference image can be obtained when the tissue region is
`not illuminated by light for use in accounting for and
`removing the inherent noise in the other images of the tissue
`region that were obtained during illumination. The body part
`can be any part of a person or an animal. For example, the
`body part can be, but is not limited to, a leg, a foot, a finger,
`or a hand. By comparing the image data for the different
`images obtained at dif