US20100257214A1 - Medical records system with dynamic avatar generator and avatar viewer - Google Patents

Medical records system with dynamic avatar generator and avatar viewer Download PDF

Info

Publication number
US20100257214A1
US20100257214A1 US12/726,959 US72695910A US2010257214A1 US 20100257214 A1 US20100257214 A1 US 20100257214A1 US 72695910 A US72695910 A US 72695910A US 2010257214 A1 US2010257214 A1 US 2010257214A1
Authority
US
United States
Prior art keywords
avatar
subject
medical
rules
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/726,959
Inventor
Luc Bessette
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US12/726,959 priority Critical patent/US20100257214A1/en
Publication of US20100257214A1 publication Critical patent/US20100257214A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H10/00ICT specially adapted for the handling or processing of patient-related medical or healthcare data
    • G16H10/60ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • G16H40/63ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation

Definitions

  • the invention relates to a system and components thereof for implementing medical records which store medical information in connection with a subject, such as a human being or an animal.
  • the system can generate an avatar of the subject and dynamically update the avatar according to the medical condition of the subject.
  • FIG. 1 is a high level block diagram of a system for implementing medical records, according to a non-limiting example of implementation of the invention
  • FIG. 2 is a high level flowchart illustrating the process for generating an avatar in connection with a subject and for updating the avatar;
  • FIG. 3 is a high level block diagram of a program executable by a computer to generate an avatar
  • FIG. 4 is a more detailed block diagram of a module of the program illustrated at FIG. 3 , for generating a personalized avatar;
  • FIG. 5 is a block diagram of a rules engine of the program module for generating a personalized avatar
  • FIG. 6 is a more detailed block diagram of a module of the program illustrated at FIG. 3 , for updating the avatar;
  • FIG. 7 is a block diagram of a rules engine of the program module for updating the avatar
  • FIG. 8 is a block diagram of a module for updating the avatar on the basis of image based and non-image based medical conditions of the subject;
  • FIG. 9 is a flow chart of the process for updating the avatar on the basis of image data obtained from the subject.
  • FIG. 10 is a block diagram of an avatar viewer module
  • FIG. 11 is a more detailed block diagram of the avatar viewer module shown in FIG. 10 .
  • avatar refers to a graphical representation of a subject which reflects the medical condition of the subject.
  • the avatar can be stored as a set of data in a machine-readable storage medium and can be represented on any suitable display device, such a two-dimensional display device, a three-dimensional display device or any other suitable display device.
  • the avatar graphically depicts medical conditions of the subject.
  • the avatar is a virtual representation of the human/animal body that is personalized according to the subject's traits or attributes and also adapted according to the medical condition of the subject.
  • a physician or any other observer can navigate the virtual representation of the body to observe the internal/external structures of the body.
  • the representation of the internal/external structures of the body can be static. Those structures can be manipulated in three dimensions or observed in cross-section by using an appropriate viewer.
  • animation techniques can simulate motion within the body or outside the body.
  • animation techniques can show a beating heart, simulate the flow of body fluids (e.g., blood) or other dynamic conditions. Motion outside the body may include, for instance, motion of limbs, such as arms, legs head, etc.
  • the components of a medical records system 10 are illustrated in FIG. 1 .
  • the system 10 has two main components, namely a medical information database 110 and a dynamic avatar generator 120 .
  • the medical information database 110 contains information of medical nature in connection with a subject, such as a human or an animal. Examples of the medical information within the database 110 can include:
  • Static information which is characterized by certain information that is inherent to the individual and is therefore not expected to change.
  • static information may include a person's name, gender, blood type, genetic information, eye color, distinguishing marks (e.g., scars, tattoos).
  • Other types of related information that could be considered static information may include:
  • a person's family medical history i.e., known conditions of their father or mother
  • information that is changeable in the longer term such as a person's current address, phone number(s), regular physician (if available), emergency contact details and/or known allergies.
  • Static information in the medical information database 110 would also include a universal or network-attributed identifier that would allow one record or file (and therefore a subject) to be distinguished from another. Use of such an identifier would allow the contents of a person's medical history to become accessible from the information database 110 .
  • Medical condition information of the subject such as a list of the subject's current or past illnesses and/or test data associated with the current and past illnesses.
  • the test data could include the test results performed on the subject such as blood tests; urine tests blood pressure tests, weight, measurements of body fat, surgeries, and results or imaging procedures such as x-rays, MRIs, CT scans and ultrasound tests, among others. The most recent results of those tests are stored in the file in addition to any previous tests performed on the subject.
  • Pharmacological data associated with the subject such as current and past drugs that have been prescribed.
  • Lifestyle information associated with the subject such as:
  • the above information may be organized within the medical information database 110 as individual records stored within the database (such as those stored within a table), or as records that are accessible to the database but are not otherwise stored within the database. Since the organization of information within databases are believed to be well known in the art, further details about the organization of the aforementioned information within the medical information database 110 need not be provided here. For additional information about medical record structures the reader may refer to the U.S. Pat. Nos. 6,775,670 and 6,263,330 the contents of which are hereby incorporated by reference.
  • the medical records system 10 includes the dynamic avatar generator 120 , which is software implemented to generate an avatar.
  • the dynamic avatar generator is a program code stored in a machine readable storage medium for execution by one or more central processing units (CPUs).
  • CPUs central processing units
  • the execution of the program code produces an avatar 130 , which is data that provides a representation of the subject and illustrates its traits and/or medical conditions.
  • the medical records system 10 may be implemented on any suitable computing platform, which may be standalone or of a distributed nature.
  • the computing platform would normally include a CPU for executing the program and a machine-readable data storage for holding the various programs and the data on which the programs operate.
  • the computing platform may be a standalone unit or of distributed nature, where different components reside at different physical locations. In this instance, the various components interoperate by communicating with one another over a data network.
  • a specific example of this arrangement is a server-client architecture, where the various databases holding the medical information reside at a certain network node and clients, which are machines on which users interact with the medical records.
  • the system 10 also implements a user interface that allows a user to access a particular medical record, modify a particular medical record and view and/or modify the avatar (depending on permission levels).
  • the user interface provides the following functionality:
  • access to the functions above may be determined on the basis of access levels, such that certain users of the system 10 can be allowed to create/modify/delete records while others are given only permissions to view the information in the record. Yet other users may be allowed to view only certain information associated with the record (such as static information), while other information associated with the subject (e.g., a subject's medical condition information) would be rendered inaccessible. In this way, the information associated with each subject within the system 10 generally, and the medical information database 110 in particular, can be protected.
  • the user interface allows a user to view the avatar 130 associated with the particular medical record.
  • a viewer module which is implemented in software, provides the user with the ability to interact with the avatar data to manipulate the data and generate the view that provides the information sought. The viewer module will be described later in greater detail.
  • FIG. 2 illustrates the general process that is implemented by the dynamic avatar generator 120 in order to create the avatar 130 .
  • the process includes two main steps.
  • the first step is the generation of the avatar for a particular subject.
  • an avatar is generated for a subject using the dynamic avatar generator 120 .
  • the program starts from a generic avatar and adapts this avatar to the subject.
  • the output of step 210 is an avatar that is tailored to the subject and represents the subject in terms of human body structure.
  • the second step of the process 220 is that of avatar updating. At that step the avatar is altered over time to reflect the medical evolution of the subject such that the avatar continues to be an accurate representation of the body of the subject. This process will be described in greater detail below.
  • FIG. 3 is a more detailed block diagram of the dynamic avatar generator 120 .
  • the dynamic avatar generator 120 has two main modules, namely an avatar personalization engine 310 and an avatar evolution engine 320 , which correspond to the two main steps of the process shown in FIG. 2 .
  • the functionality of the avatar personalization engine 310 is discussed below with regards to the generation of a new avatar, which is associated with step 210 .
  • the functionality of the avatar evolution engine 320 will be discussed later in the context of updating the avatar, which occurs at step 220 .
  • the avatar personalization engine 310 is used to customize the avatar 130 in certain ways so that it can represent its corresponding subject more realistically.
  • the engine 310 can be used to personalize both an avatar's external appearance, as well as adjust its internal organ structure so that the avatar 130 is as faithful a representation of its corresponding subject as possible.
  • the avatar personalization engine 310 allows the generic avatar 130 to be personalized in two (2) ways, namely an external personalization and an internal personalization. External personalization involves adjusting the appearance and structure of the avatar 130 so that it represents the appearance of its corresponding subject. To provide this control, the avatar personalization engine 310 provides tools to the user via the user interface to control all aspects of the avatar's 130 external appearance.
  • an avatar's external personalization may be manually configured, such setting a particular eye color or hair texture (e.g., curly or straight) for the avatar 130 .
  • Other aspects of external personalization for the avatar 130 may be automatically configured by the personalization engine 310 based on a user's choices, such as those based on a chosen sex (i.e., whether the subject is male or female). For example, indicating that a subject is male allows the avatar personalization engine 310 to include male reproductive organs within the appearance of the avatar 130 .
  • such indications allow the personalization engine 310 to pre-configure a number of aspects of an avatar's appearance simultaneously that may save a user time and effort.
  • the avatar personalization engine 310 may provide the ability to ‘import’ a photograph of the corresponding subject (which may be in two- or three-dimensions) so that this photograph may be used to further personalize the avatar.
  • the avatar personalization engine 310 could apply a frontal photograph of the face of the subject to the “face” of the avatar 130 such that the avatar's face resembles that of its subject. This could be done either by simply wrapping the photograph as a texture to the default face of the avatar, or by extracting biometric information from the photograph such that biometric features in the face of the avatar 130 would be adjusted in a similar fashion.
  • the avatar personalization engine 310 could use a two- or three-dimensional photograph of the subject's body in order to apply similar body measurements to the appendages of the avatar 130 .
  • the engine 310 could extract biometric information about the relative length of the arms and/or legs to the torso of the subject in order that the same relative lengths would be applied to the avatar 130 .
  • the result of the external personalization process is the production of an instance of the avatar 130 whose appearance resembles that of its corresponding subject. While certain means for such personalization have been described above, it will be appreciated that other ways of personalizing the external appearance of an avatar exist and would fall within the scope of the invention.
  • the avatar personalization engine 310 also allows the internal organs and systems (e.g., veins and arteries in the circulatory systems) comprised in the avatar 130 to be customized.
  • every avatar 130 is created with a generic set of individual organs and systems for their chosen sex, which are supposed to correspond to the subject's set of internal organs and systems.
  • This generic set of organs and systems are also controlled by a set of rules and conditions that define how these organs are supposed to work by default.
  • the avatar personalization engine 310 can be used to more closely match the organs and systems of the avatar 130 to those of its corresponding subject.
  • the default ‘heart rate’ for an avatar representing a 40-year old male may be defined as 80 beats per minute, but a man's default heart rate is actually recorded at beats/minute.
  • the personalization engine 310 sets the heart rate of the man's avatar to 95 beats/minute as well.
  • Use of the avatar personalization engine 310 to adjust or customize the avatar 130 may be initiated in several ways, including:
  • the manual adjustment which may be based on a person's input, namely the person opening the medical record or creating the personalized avatar.
  • the manual adjustment may include for different internal body structures a list of possible choices and the person simply chooses the option that suits the subject best;
  • automatic adjustment which may be based on existing information in medical records and/or photos or other data that represents the subject;
  • biometric adjustment which may be based on a scan of the person's body such as from a CT scan X-rays, MRIs or others.
  • automatic and/or biometric adjustments of the avatar 130 may be implemented by a separate image processing software module that is initiated by the avatar personalization engine 310 .
  • the software module may process the image data (which may be two-dimensional, such as in X-ray images or three-dimensional, such as in CT scans) in order to detect certain salient features of the scanned internal structures in the image and then apply those features to the avatar 130 .
  • the avatar personalization engine 310 receives an X-ray image of a bone and surrounding tissue for a subject.
  • the engine 310 may submit this image to the image processing software module in order to extract measurements so that a three-dimensional model of the bone and surrounding tissue (e.g., muscles) can be replicated in the avatar.
  • the software module may process the image in order to identify certain features of the bone, such as its dimensions, that may be identified by observing and identifying differences in the gray-scale gradient between the bone and surrounding tissue that exceed a certain known value. By identifying the dimensions of the bone from the two-dimensional X-ray image, a three-dimensional model of the corresponding bone can be created and applied to the avatar 130 for the subject. Similar processes may be used by the image processing software module to observe and identify different tissues (e.g., muscle tissue versus tissue for veins or arteries) within the surrounding tissue in order that three-dimensional models of such tissues can be generated.
  • tissue e.g., muscle tissue versus tissue for veins or arteries
  • the image processing software module used by the avatar personalization engine 310 may also process three-dimensional data (such as that supplied by a CT scan) in a similar manner.
  • the personalization step 210 may be a one-time processing operation or a continuous process that refines the avatar 130 over time.
  • the initial medical information available on the subject may be limited and may not include a complete set of medical data to personalize every structure of the body.
  • the avatar 130 may only be partially personalized by the engine 310 , and body features and structures for which no medical information is available from the subject would not be modified from their generic or default version.
  • new medical information becomes available (such as an X-ray image of a bone that was never imaged before)
  • that information can be used by the avatar personalization engine 310 to further personalize the avatar 130 by altering the generic version of the bone to acquire the features observed in the X-ray.
  • the avatar personalization engine 310 has the ability to apply certain exceptions to the appearance and/or internal physiology of the avatar 130 . For example, assume that a 20-year old male soldier has lost his right leg below the knee. To make his avatar as representative as possible, the engine 310 may be used to remove his right leg and foot from the avatar's external appearance. In certain cases, the avatar may be provided with a prosthetic leg and foot that correspond to the prosthetics actually used by the male soldier.
  • the internal physiology of the avatar's right leg may be further adjusted by the avatar personalization engine 310 such that the bones, veins, arteries and nerve endings terminate at the same point as they do in the soldier's real leg.
  • Such customization to the avatar may be initiated by and/or based on X-ray or CT scans of the area in question.
  • FIG. 4 is a yet more detailed block diagram of the avatar personalization engine 310 .
  • the avatar personalization engine 310 operates on the basis of a set of personalization rules that condition a set of input data to create a personalized avatar.
  • the input conditions can be represented by a Human Anatomy and Composition Representation database 410 (referred to as the HACR database hereafter).
  • the contents of the HACR database 410 include the input conditions that anatomically define the external appearance and/or internal physiology of each generated instance of the avatar 130 .
  • the HACR database 410 may be seen as providing a similar function as that typically provided by human or animal DNA, but at a much higher level, in that the database 410 provides a default template for the composition and construction of each instance of the avatar 130 .
  • the contents of the HACR database 410 are structured and organized according to a Body Markup Language (BMR), which is a language that expresses body (human or animal) structures.
  • BMR Body Markup Language
  • a BML functions by associating a certain structure of the body with a tag.
  • Each tag defines the characteristics of the body structure, such as how the body structure would appear when it is viewed and how it relates to other body structures. Therefore, a BML representation of the body requires breaking down the body into individual structures and then associating each structure to a tag.
  • the structures (and their associated tags) described above define an implicit anatomical and physiological taxonomy of an animal or human body whose granularity in terms of individual structures may vary depending on the application. For example, while single cells could be considered as individual structures within the taxonomy of the tagging language, given the huge number of cells in a body, exceedingly large computational resources would be required to express the body structure at such a fine level of detail. Conversely at the other end of the taxonomy, body structures can be simplified to individual systems, such as where the entire urinary system or the respiratory system can be considered as a single discrete structure.
  • Each individual structure can be represented as image data stored in a machine readable storage medium.
  • the image data can be in any suitable format without departing from the spirit of the invention.
  • the degree of image detail for each individual structure can vary depending on the intended application.
  • the image data for a structure may be as simple as including a two-dimensional image of the structure, such as an image extracted from an X-ray scan.
  • the image data can include a three-dimensional image of the structure, such that during visualization the image can be manipulated so that it can be seen from different perspectives.
  • Another possibility is to provide a structure that can be represented by a three-dimensional modeling program on the basis of a three-dimensional mesh.
  • the mesh can be resized, stretched or otherwise modified to change the shape of the basic organ.
  • the three-dimensional modeler also can include a texture-mapping feature that can apply textures onto the mesh.
  • the three-dimensional modeler can be used to generate a three dimensional image of the outside of the structure but also can be used to generate a complete three dimensional representation of the entire structure, showing its outside surface and also its internal features as well.
  • this form of representation could be used to show the internal structure of the human heart, therefore allowing a user to see the outside of the heart, manipulate the heart to see it from different angles, take virtual ‘slices’ (cross-sections) of the heart to expose the inside structure at a certain point or ‘fly through’ the heart in order to review its external or internal structure.
  • Yet another possibility is to provide image data that actually contains several different representations of the organ, which may be two-dimensional, three-dimensional or could be represented by a three-dimensional modeling program.
  • the various representations of the organ could be individually analyzed and then combined to form a single organ based on observed overlaps between the different representations or prior knowledge of the structure of the organ.
  • Each structure is further associated with a tag that contains instructions about the manner in which the image data behaves. Examples of such instructions include:
  • the information within the HACR database 410 may be subjected one or more rules in a set of personalization rules 420 .
  • the rules 420 define certain conditions or settings that adjust the appearance or internal physiology of the avatar 130 in concordance with that observed in the corresponding subject.
  • the information within the HACR database 410 may be subjected one or more rules in a set of personalization rules 420 .
  • the rules 420 determine how the generic avatar will be altered to match the subject.
  • the personalization rules include logic that alters the image data associated with the respective body structures. That logic is embedded in the tags of the respective structures such that the behavior of the image data corresponding to the structures changes as desired.
  • the image alterations during a personalization process of the generic avatar are designed to perform the following, including among others, aging, changes to corporeal traits, changes based on gender or race, as well as possible exceptions. Further information about these alterations defined by the personalization rules are provided below.
  • Aging (or age adjustment) rules refer to adjustment rules that are intended to adjust the visual appearance of the set of structures comprising the avatar 130 so that they match the age of the subject.
  • a set of age adjustment rules exist, where different aging rules apply to different structures, as different structures are affected in a different way as a result of aging.
  • Each age adjustment rule models the effect of aging on a structure and in particular on how a structure appears.
  • the model which as indicated earlier, may be specific to an individual structure or may affect a set of structures, can be based on empirical observation on the effect of aging on body structures. For example, in the case of human bones, aging can affect the bone dimensions and its density. As a person ages, his or her bones are likely to shrink slightly and also become more porous.
  • an aging rule will typically include logic that changes the image data such that the image of the bone is resized as a function of age.
  • the older the subject the smaller his or her bones will appear.
  • the degree of re-sizing can be derived from medical knowledge and observation and would generally be known to those skilled in the art.
  • Another age adjustment rule for human bones may be used to depict changes to bone porosity with age.
  • pores are created in the image (either at random positions or at predetermined positions), where the number of pores and their size is dependent on the age of the subject. As a result, the older the subject, the higher the number of pores and the larger their size will be.
  • an aging rule may relate to the pigmentation, color and texture of the skin.
  • the age adjustment rule associated with these body structures define a texture and color model that is age-dependent.
  • the texture and color gradations can be based on empirical observations that mimic how the skin ages.
  • the texture and color models will render the image of these structures in a way that will realistically mimic the skin of an older person on the avatar 130 .
  • the model may control the rendering of the surface of the skin, such that the skin looks rougher and may have small dark dots randomly distributed.
  • an age adjustment rule could be a rule that affects the appearance of the prostate gland. As is generally well known, the size of the prostate often becomes enlarged with age. The age adjustment rule would therefore be designed to alter the size (and possibly shape) of the prostate such that it becomes larger with age.
  • an aging rule may be one associated with gums. It is well known that as a person ages, his or her gums recede. Accordingly, the model implemented by the age adjustment rule would be designed to alter the image data of the gums such that the gums appear as receding, where the amount of receding is dependent on age.
  • age adjustment rules can also be provided that alter certain kinetic functions which are known to be age-dependent. For instance, age typically affects the range of motion at a joint, such as the knee or elbow. To model these effects, an aging rule may be implemented that when the avatar 130 displays movement at those joints, the motion is restricted to a range that is age dependent. As a result, the avatars of higher aged subject would have a lower range of motion for the affected joints and related structures.
  • simulation of other motions can be conditioned in a similar way.
  • the general heart beat rate for the avatar 130 may be lowered as age increases to reflect known medical knowledge about the relationship between a person's heart rate and his or her age.
  • the personalization rules engine may also include the following:
  • FIG. 6 is a yet more detailed block diagram of the avatar evolution engine 320 .
  • the avatar evolution engine 320 operates on the basis of a set of ‘evolution’ rules that condition a set of input data to update an avatar from a prior state to a new state.
  • the starting point of the updating process is the personalized avatar.
  • the personalized avatar therefore is altered progressively by the updating engine such as the avatar continues to represent the subject as the body of the subject evolves over time and changes due to aging and medical conditions.
  • the changes to the avatar made by the updating rules engine can include progressive changes such as those due to aging and discrete changes resulting from specific medical conditions encountered.
  • FIG. 6 The set of ‘evolution’ rules that condition the input data in order to update an avatar from a prior state to a new state are represented in FIG. 6 by an updating rules engine 620 .
  • FIG. 7 shows various categories of rules that may be included within the set of evolution rules represented by the engine 620 , which could include among others:
  • one or more of the rules (and in particular, the observed medical condition rules) in the updating rules engine 620 described above may originate from the medical information database 110 .
  • the genetic rules may originate from the genetic profile of the corresponding subject, which may be stored in the database 110 .
  • the aging rules in the updating rules engine 620 may be updated or altered based on contributions from the other rule categories.
  • the geographic and/or demographic group categories may cause an adjustment in the aging rules that causes the avatar 130 to age faster or slower than would otherwise be expected.
  • the aging rules for a Chinese male laborer who lives in or around central Shanghai, China and smokes at least two (2) packs of cigarettes a day would likely cause this subject's avatar to age more quickly than otherwise.
  • the identification of certain genetic conditions in the genetic profile of a subject that confer improved resistance to certain diseases that are more common to a particular demographic group (e.g., resistance to heart disease in people 50 and above) that may be expressed in the genetic rules may cause the avatar 130 to age more slowly than would otherwise be expected.
  • the various rules within the updating rule engine 620 only govern the evolution of the avatar 130 between two states separated by time, namely a first, earlier state and a second, later state. It will be appreciated that such changes may or may not relate to the actual physiological evolution of the avatar's corresponding subject. In cases where the evolved state of the avatar 130 differs from that of its corresponding subject, the avatar 130 may be further updated based on observed medical conditions.
  • FIG. 8 illustrates a non-limiting method by which the updating rule engine 620 may update the avatar 130 between a first prior state and a second, more current state based on observed medical conditions.
  • This figure includes two (2) sets of data, namely a medical non-image dataset 810 and a medical image dataset 820 .
  • the datasets 810 and 820 are presented here as separate entities, this is done for the sake of illustration. In reality, both of these datasets are quite likely to reside together in the medical information database 110 .
  • the contents of the medical non-image dataset 810 typically contain medical information for the subject that is non-visual in nature, such as numeric test results, observation notes by medical personnel and/or biopsy reports, among others. Moreover, contents of this dataset may be linked to certain aspects of the HACR database 410 , such as tagged content within the structures component 412 and/or internal kinetics component 414 . For example, a test showing the blood pressure and flow through specific arteries in the cardiovascular system may be used to model blood flow in the avatar 130 .
  • the contents of the medical image dataset 820 include medical information that is visual in nature, such as X-ray images, photographs taken from a biopsy and/or CT-scan related data and ultrasound observations among others.
  • contents of this dataset may be linked to or associated with the HACR database 410 in order to provide the images used for tagged structures within the structures component 412 .
  • an X-ray of a leg bone and surrounding tissue may be associated with the tagged structure that defines how the leg bone and surrounding tissue in the avatar 130 is represented.
  • the avatar evolution engine 320 may monitor the contents of the datasets 810 and 820 in order that it can become aware of any new information that is added to these datasets from observation of the subject. Alternatively, the engine 320 may be advised of the addition of new data to the datasets 810 and 820 only at the time when the avatar 130 is to be updated.
  • the avatar evolution engine 320 becomes aware of new information in the datasets 810 and 820 , it can use this information to update the observed medical condition rules components of the updating rules engine 620 in order to update the avatar 130 in a similar fashion.
  • new information within the medical non-image dataset 810 could be used to update a set of non-image based updating rules 815 that may be included within the observed medical condition rules category in the updating rules engine 620 .
  • new information within the medical image dataset 820 could be used to update a set of image-based updating rules 825 , which may also be included within the observed medical condition rules category in the updating rules engine 620 .
  • a subject suffers a fall and that their brain is subjected to a CT scan to ensure that they are not suffering from a condition, such as a concussion or brain edema.
  • the data generated by the CT scan of the subject's brain is stored within the medical information database 110 and becomes part of the medical image dataset 820 as a result.
  • the addition of this data to the dataset 820 may trigger the avatar evolution engine 320 to review and revise the updating rules engine 620 based on this new information.
  • the engine 620 may use this data to update the observed medical condition rules category for the brain to update the previous image associated with the tagged brain entry in the structures component 410 (which would likely have been taken before the fall) with a new image updated to take into account the information contained in the CT scan data.
  • the avatar evolution engine 320 now has two separate brain images from the subject, it can evaluate changes between the images in order to update the brain represented in the avatar 130 in the same way. This can be done by updating the observed medical condition category of the updating rules engine 620 to account for the new image, which may involve adjusting the image modifier information for the tag associated with the brain structure.
  • FIG. 9 shows a flowchart that illustrates a non-limiting process by which information in the previously mentioned medical image dataset 820 (and/or the medical information database 110 as a whole) could be used to update the avatar 130 .
  • image data is processed to identify the particular structure (i.e., body part or organ of the avatar) to which the image applies.
  • This data may be processed by the avatar evolution engine 320 , by the updating rules engine 620 or by an image processing software module that is likely similar (if not identical) to discussed in the context of avatar personalization.
  • the structure or body part to which the image applies may be included within the image data.
  • an image taken of a femur bone may include metadata (which may be based on BML) that may indicate the image was of a left femur bone.
  • image-related information that might be provided within the image data may include the angle at which the image was taken, the device used to generate the image and/or an indication as to why the image was generated (e.g., as part of a standard checkup or as a result of certain trauma).
  • the process may proceed to the next step immediately. However, if this information is missing from or is not included with the image data, it may be extracted at this point by analyzing the medical image and comparing any features identified within it against known body structures and/or structures within the subject's body in particular.
  • the size and shape of the bone may be compared to those found in existing images and/or models to see which of these produce the closest match.
  • the identified bone may be compared against bones within the avatar 130 and/or medical image dataset 820 , and more specifically, images that contain bones with a similar size, shape and/or orientation.
  • the image processing and/or pattern matching may be performed against bones associated with the avatar 130 of the subject, against bones associated with the medical image dataset 820 or against bones known to be in images stored within the medical information database 110 . This can increase the likelihood that a bone that is captured within an image will be matched correctly to its corresponding structure.
  • relevant features are extracted from the image data.
  • the image is processed to identify relevant features in the structure, which may include among others:
  • the process by which relevant features may be identified may include comparing the structure within in the current image with an image of the structure taken at a prior state, which may be stored within the medical image dataset 820 .
  • an X-ray image of a subject's femur may be compared against earlier X-ray images of the same bone to identify any changes that have taken place.
  • steps 910 and 920 in FIG. 9 are shown in sequential order, the processing of the image that occurs in these steps may also be performed more or less simultaneously. Therefore, while the image is being processed to identify its corresponding structure (step 910 ), it may also be processed simultaneously to identify relevant features (step 920 ).
  • the result of the previous step was the identification of relevant features for the structure based on image data from the subject.
  • the avatar 130 In order to ensure the avatar 130 reflects the state of its corresponding subject, the avatar must be updated in the same way.
  • the avatar 130 is updated to include the same relevant features as were identified during the previous step.
  • This update to the avatar 130 is typically done by the avatar evolution engine 320 via the updating rules engine 620 . More specifically, the update may be performed by the engine 320 using the updated non-image based input rules 815 and image-based input rules 825 of the observed medical conditions rule category residing within the engine 620 .
  • step 930 the avatar 130 will have been updated to reflect the most current medical condition of its corresponding subject. This process prepares the avatar 130 for viewing by medical personnel in order to diagnose and/or treat medical conditions affecting the subject.
  • FIG. 11 is a block diagram of an image viewer that can used for viewing the updated avatar in its entirety of components thereof.
  • the image viewer is associated with the user interface of the medical record and a user can invoke the viewer from the user interface control.
  • the viewer includes a body structures module 1020 that allows the user to select the particular structure or set of structures for display. For instance, the user can select a single structure to be shown, such as a bone or an organ, say the heart.
  • the viewer can provide navigational tools allowing the user to rotate the image such that it can be seen from different perspectives, create slices to see the inside of the structure, among others. In the specific example shown, a slice through the entire body of the avatar is illustrated.
  • the viewer allows the user to display a series of structures that are related to one another, either by virtue of physical relation or functional relation.
  • the viewer module also has a kinetics viewer that can show animation of selected structures. For instance the kinetics viewer can animate a joint and depict how the various bones move, simulate the beating of the heart, simulate the blood flow through a certain organ, etc.

Abstract

A medical records system, including a data processing apparatus including a CPU and a computer readable storage medium encoded with a program for execution by the CPU. The program processes medical information in connection with a subject to generate an avatar of the subject which reflects the medical status of the subject.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application claims benefit of priority of U.S. Provisional Patent Application No. 61/161,299, filed Mar. 18, 2009, which is incorporated herein by reference.
  • FIELD OF THE INVENTION
  • The invention relates to a system and components thereof for implementing medical records which store medical information in connection with a subject, such as a human being or an animal. The system can generate an avatar of the subject and dynamically update the avatar according to the medical condition of the subject.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • A detailed description of examples of implementation of the present invention is provided hereinbelow with reference to the following drawings, in which:
  • FIG. 1 is a high level block diagram of a system for implementing medical records, according to a non-limiting example of implementation of the invention;
  • FIG. 2 is a high level flowchart illustrating the process for generating an avatar in connection with a subject and for updating the avatar;
  • FIG. 3 is a high level block diagram of a program executable by a computer to generate an avatar;
  • FIG. 4 is a more detailed block diagram of a module of the program illustrated at FIG. 3, for generating a personalized avatar;
  • FIG. 5 is a block diagram of a rules engine of the program module for generating a personalized avatar;
  • FIG. 6 is a more detailed block diagram of a module of the program illustrated at FIG. 3, for updating the avatar;
  • FIG. 7 is a block diagram of a rules engine of the program module for updating the avatar;
  • FIG. 8 is a block diagram of a module for updating the avatar on the basis of image based and non-image based medical conditions of the subject;
  • FIG. 9 is a flow chart of the process for updating the avatar on the basis of image data obtained from the subject;
  • FIG. 10 is a block diagram of an avatar viewer module;
  • FIG. 11 is a more detailed block diagram of the avatar viewer module shown in FIG. 10.
  • In the drawings, embodiments of the invention are illustrated by way of example. It is to be expressly understood that the description and drawings are only for purposes of illustration and as an aid to understanding, and are not intended to be a definition of the limits of the invention.
  • DETAILED DESCRIPTION
  • For the purposes of the present specification, the expression “avatar” refers to a graphical representation of a subject which reflects the medical condition of the subject. The avatar can be stored as a set of data in a machine-readable storage medium and can be represented on any suitable display device, such a two-dimensional display device, a three-dimensional display device or any other suitable display device.
  • The avatar graphically depicts medical conditions of the subject. In one specific and non-limiting example, the avatar is a virtual representation of the human/animal body that is personalized according to the subject's traits or attributes and also adapted according to the medical condition of the subject. A physician or any other observer can navigate the virtual representation of the body to observe the internal/external structures of the body. The representation of the internal/external structures of the body can be static. Those structures can be manipulated in three dimensions or observed in cross-section by using an appropriate viewer. It is also possible to use animation techniques to simulate motion within the body or outside the body. For example, animation techniques can show a beating heart, simulate the flow of body fluids (e.g., blood) or other dynamic conditions. Motion outside the body may include, for instance, motion of limbs, such as arms, legs head, etc.
  • The components of a medical records system 10 are illustrated in FIG. 1. The system 10 has two main components, namely a medical information database 110 and a dynamic avatar generator 120. The medical information database 110 contains information of medical nature in connection with a subject, such as a human or an animal. Examples of the medical information within the database 110 can include:
  • 1) Static information, which is characterized by certain information that is inherent to the individual and is therefore not expected to change. Examples of static information may include a person's name, gender, blood type, genetic information, eye color, distinguishing marks (e.g., scars, tattoos). Other types of related information that could be considered static information may include:
  • a person's family medical history (i.e., known conditions of their father or mother);
  • information that is changeable in the longer term, such as a person's current address, phone number(s), regular physician (if available), emergency contact details and/or known allergies.
  • Static information in the medical information database 110 would also include a universal or network-attributed identifier that would allow one record or file (and therefore a subject) to be distinguished from another. Use of such an identifier would allow the contents of a person's medical history to become accessible from the information database 110.
  • (2) Medical condition information of the subject, such as a list of the subject's current or past illnesses and/or test data associated with the current and past illnesses. The test data could include the test results performed on the subject such as blood tests; urine tests blood pressure tests, weight, measurements of body fat, surgeries, and results or imaging procedures such as x-rays, MRIs, CT scans and ultrasound tests, among others. The most recent results of those tests are stored in the file in addition to any previous tests performed on the subject.
    (3) Pharmacological data associated with the subject, such as current and past drugs that have been prescribed.
    (4) Lifestyle information associated with the subject, such as:
      • 1. whether the subject is a smoker or non-smoker;
      • 2. the level of the subject's Physical fitness (e.g., super fit, medium fit or not fit);
      • 3. the amount of body fat (lean/average/obese), which may be determined through a measurement of the subject's BMI index;
  • It will appreciated that the above information may be organized within the medical information database 110 as individual records stored within the database (such as those stored within a table), or as records that are accessible to the database but are not otherwise stored within the database. Since the organization of information within databases are believed to be well known in the art, further details about the organization of the aforementioned information within the medical information database 110 need not be provided here. For additional information about medical record structures the reader may refer to the U.S. Pat. Nos. 6,775,670 and 6,263,330 the contents of which are hereby incorporated by reference.
  • In addition to the medical information database 110, the medical records system 10 includes the dynamic avatar generator 120, which is software implemented to generate an avatar. The dynamic avatar generator is a program code stored in a machine readable storage medium for execution by one or more central processing units (CPUs). The execution of the program code produces an avatar 130, which is data that provides a representation of the subject and illustrates its traits and/or medical conditions.
  • The medical records system 10 may be implemented on any suitable computing platform, which may be standalone or of a distributed nature.
  • The computing platform would normally include a CPU for executing the program and a machine-readable data storage for holding the various programs and the data on which the programs operate. The computing platform may be a standalone unit or of distributed nature, where different components reside at different physical locations. In this instance, the various components interoperate by communicating with one another over a data network. A specific example of this arrangement is a server-client architecture, where the various databases holding the medical information reside at a certain network node and clients, which are machines on which users interact with the medical records.
  • To allow a user to interact with the medical records system 10, the system 10 also implements a user interface that allows a user to access a particular medical record, modify a particular medical record and view and/or modify the avatar (depending on permission levels). In particular, the user interface provides the following functionality:
      • 1. Create a medical record in connection with a subject
      • 2. View an existing medical record in connection with a certain subject;
      • 3. Modify an existing medical record in connection with a certain subject such as entering data regarding a medical test performed on the subject;
      • 4. Delete a medical record.
  • For security purposes, access to the functions above may be determined on the basis of access levels, such that certain users of the system 10 can be allowed to create/modify/delete records while others are given only permissions to view the information in the record. Yet other users may be allowed to view only certain information associated with the record (such as static information), while other information associated with the subject (e.g., a subject's medical condition information) would be rendered inaccessible. In this way, the information associated with each subject within the system 10 generally, and the medical information database 110 in particular, can be protected.
  • The user interface allows a user to view the avatar 130 associated with the particular medical record. A viewer module, which is implemented in software, provides the user with the ability to interact with the avatar data to manipulate the data and generate the view that provides the information sought. The viewer module will be described later in greater detail.
  • FIG. 2 illustrates the general process that is implemented by the dynamic avatar generator 120 in order to create the avatar 130. The process includes two main steps. The first step is the generation of the avatar for a particular subject. At step 210, an avatar is generated for a subject using the dynamic avatar generator 120. In short, at this step, the program starts from a generic avatar and adapts this avatar to the subject. The output of step 210 is an avatar that is tailored to the subject and represents the subject in terms of human body structure.
  • The second step of the process 220 is that of avatar updating. At that step the avatar is altered over time to reflect the medical evolution of the subject such that the avatar continues to be an accurate representation of the body of the subject. This process will be described in greater detail below.
  • FIG. 3 is a more detailed block diagram of the dynamic avatar generator 120. The dynamic avatar generator 120 has two main modules, namely an avatar personalization engine 310 and an avatar evolution engine 320, which correspond to the two main steps of the process shown in FIG. 2.
  • The functionality of the avatar personalization engine 310 is discussed below with regards to the generation of a new avatar, which is associated with step 210. The functionality of the avatar evolution engine 320 will be discussed later in the context of updating the avatar, which occurs at step 220.
  • The avatar personalization engine 310 is used to customize the avatar 130 in certain ways so that it can represent its corresponding subject more realistically. The engine 310 can be used to personalize both an avatar's external appearance, as well as adjust its internal organ structure so that the avatar 130 is as faithful a representation of its corresponding subject as possible.
  • Use of the avatar personalization engine 310 allows the generic avatar 130 to be personalized in two (2) ways, namely an external personalization and an internal personalization. External personalization involves adjusting the appearance and structure of the avatar 130 so that it represents the appearance of its corresponding subject. To provide this control, the avatar personalization engine 310 provides tools to the user via the user interface to control all aspects of the avatar's 130 external appearance.
  • Certain aspects of an avatar's external personalization may be manually configured, such setting a particular eye color or hair texture (e.g., curly or straight) for the avatar 130. Other aspects of external personalization for the avatar 130 may be automatically configured by the personalization engine 310 based on a user's choices, such as those based on a chosen sex (i.e., whether the subject is male or female). For example, indicating that a subject is male allows the avatar personalization engine 310 to include male reproductive organs within the appearance of the avatar 130. Advantageously, such indications allow the personalization engine 310 to pre-configure a number of aspects of an avatar's appearance simultaneously that may save a user time and effort.
  • Although the use of indications (such as indicating the sex of the subject) in order to pre-configure a number of aspects of the avatar's appearance can be helpful, further personalizing the avatar so that it resembles the subject may require considerable time. To reduce the amount of time required, the avatar personalization engine 310 may provide the ability to ‘import’ a photograph of the corresponding subject (which may be in two- or three-dimensions) so that this photograph may be used to further personalize the avatar.
  • For example, the avatar personalization engine 310 could apply a frontal photograph of the face of the subject to the “face” of the avatar 130 such that the avatar's face resembles that of its subject. This could be done either by simply wrapping the photograph as a texture to the default face of the avatar, or by extracting biometric information from the photograph such that biometric features in the face of the avatar 130 would be adjusted in a similar fashion.
  • Similarly, the avatar personalization engine 310 could use a two- or three-dimensional photograph of the subject's body in order to apply similar body measurements to the appendages of the avatar 130. For example, the engine 310 could extract biometric information about the relative length of the arms and/or legs to the torso of the subject in order that the same relative lengths would be applied to the avatar 130.
  • The result of the external personalization process is the production of an instance of the avatar 130 whose appearance resembles that of its corresponding subject. While certain means for such personalization have been described above, it will be appreciated that other ways of personalizing the external appearance of an avatar exist and would fall within the scope of the invention.
  • Similarly, the avatar personalization engine 310 also allows the internal organs and systems (e.g., veins and arteries in the circulatory systems) comprised in the avatar 130 to be customized. By default, every avatar 130 is created with a generic set of individual organs and systems for their chosen sex, which are supposed to correspond to the subject's set of internal organs and systems. This generic set of organs and systems are also controlled by a set of rules and conditions that define how these organs are supposed to work by default.
  • Because no subject's organs or systems will be exactly same as this ‘generic’ set, the avatar personalization engine 310 can be used to more closely match the organs and systems of the avatar 130 to those of its corresponding subject.
  • For example, the default ‘heart rate’ for an avatar representing a 40-year old male may be defined as 80 beats per minute, but a man's default heart rate is actually recorded at beats/minute. To accommodate this difference, the personalization engine 310 sets the heart rate of the man's avatar to 95 beats/minute as well. Those skilled in the art will appreciate that other adjustments to the internal physiology of the avatar 130 may be made in a similar manner.
  • Use of the avatar personalization engine 310 to adjust or customize the avatar 130 may be initiated in several ways, including:
  • manual adjustment, which may be based on a person's input, namely the person opening the medical record or creating the personalized avatar. The manual adjustment may include for different internal body structures a list of possible choices and the person simply chooses the option that suits the subject best;
  • automatic adjustment, which may be based on existing information in medical records and/or photos or other data that represents the subject; and/or
  • biometric adjustment, which may be based on a scan of the person's body such as from a CT scan X-rays, MRIs or others.
  • It is worth noting that automatic and/or biometric adjustments of the avatar 130 may be implemented by a separate image processing software module that is initiated by the avatar personalization engine 310. Upon such initiation, the software module may process the image data (which may be two-dimensional, such as in X-ray images or three-dimensional, such as in CT scans) in order to detect certain salient features of the scanned internal structures in the image and then apply those features to the avatar 130.
  • For example, assume that the avatar personalization engine 310 receives an X-ray image of a bone and surrounding tissue for a subject. The engine 310 may submit this image to the image processing software module in order to extract measurements so that a three-dimensional model of the bone and surrounding tissue (e.g., muscles) can be replicated in the avatar. The software module may process the image in order to identify certain features of the bone, such as its dimensions, that may be identified by observing and identifying differences in the gray-scale gradient between the bone and surrounding tissue that exceed a certain known value. By identifying the dimensions of the bone from the two-dimensional X-ray image, a three-dimensional model of the corresponding bone can be created and applied to the avatar 130 for the subject. Similar processes may be used by the image processing software module to observe and identify different tissues (e.g., muscle tissue versus tissue for veins or arteries) within the surrounding tissue in order that three-dimensional models of such tissues can be generated.
  • Although the above example used a two-dimensional X-ray as the basis for generating a three-dimensional model, it will be appreciated that the image processing software module used by the avatar personalization engine 310 may also process three-dimensional data (such as that supplied by a CT scan) in a similar manner.
  • Note that the personalization step 210, as shown and described above, may be a one-time processing operation or a continuous process that refines the avatar 130 over time. For example, the initial medical information available on the subject may be limited and may not include a complete set of medical data to personalize every structure of the body. Accordingly, in such instances, the avatar 130 may only be partially personalized by the engine 310, and body features and structures for which no medical information is available from the subject would not be modified from their generic or default version. However, as new medical information becomes available (such as an X-ray image of a bone that was never imaged before), that information can be used by the avatar personalization engine 310 to further personalize the avatar 130 by altering the generic version of the bone to acquire the features observed in the X-ray.
  • It will also be appreciated that the avatar personalization engine 310 has the ability to apply certain exceptions to the appearance and/or internal physiology of the avatar 130. For example, assume that a 20-year old male soldier has lost his right leg below the knee. To make his avatar as representative as possible, the engine 310 may be used to remove his right leg and foot from the avatar's external appearance. In certain cases, the avatar may be provided with a prosthetic leg and foot that correspond to the prosthetics actually used by the male soldier.
  • In addition, the internal physiology of the avatar's right leg may be further adjusted by the avatar personalization engine 310 such that the bones, veins, arteries and nerve endings terminate at the same point as they do in the soldier's real leg. Such customization to the avatar may be initiated by and/or based on X-ray or CT scans of the area in question.
  • FIG. 4 is a yet more detailed block diagram of the avatar personalization engine 310. The avatar personalization engine 310 operates on the basis of a set of personalization rules that condition a set of input data to create a personalized avatar. The input conditions can be represented by a Human Anatomy and Composition Representation database 410 (referred to as the HACR database hereafter).
  • The contents of the HACR database 410 include the input conditions that anatomically define the external appearance and/or internal physiology of each generated instance of the avatar 130. In this respect, the HACR database 410 may be seen as providing a similar function as that typically provided by human or animal DNA, but at a much higher level, in that the database 410 provides a default template for the composition and construction of each instance of the avatar 130.
  • The contents of the HACR database 410 are structured and organized according to a Body Markup Language (BMR), which is a language that expresses body (human or animal) structures. A BML functions by associating a certain structure of the body with a tag. Each tag defines the characteristics of the body structure, such as how the body structure would appear when it is viewed and how it relates to other body structures. Therefore, a BML representation of the body requires breaking down the body into individual structures and then associating each structure to a tag.
  • Examples of individual structures that would likely be found in the BML include:
      • 1. Skeletal structure—where each bone of the skeleton (for the sake of the description, assume a human skeleton with 206 bones) can be a discrete structure;
      • 2. Respiratory system—where each component of the respiratory system (e.g., airways, lungs and respiratory muscles) can be a discrete structure;
      • 3. Circulatory system—where each component of the circulatory system (e.g., the blood distribution network; (2) blood pumping system (heart) and lymph distribution network) is a discrete structure;
      • 4. Muscular system—where each individual muscle (e.g., bicep and tricep in an arm) is a discrete structure;
      • 5. Nervous system—where each component of the central nervous system network and the peripheral nervous system network are discrete structures (e.g., spinal cord, sciatic nerve);
      • 6. Digestive system—where each component of the digestive system (e.g., mouth, teeth, esophagus, stomach, small intestine and large intestine) is a discrete structure;
      • 7. Urinary system—where each component of the urinary system (e.g., kidneys, bladder, urethra and sphincter muscles) is a discrete structure;
      • 8. Reproductive system—where each component of the reproductive system (e.g., the genitalia (distinguished on the basis of gender), gamete producing gonads for males and ovaries for females) is a discrete structure.
        Note that the above are examples only.
  • It is worth noting that the structures (and their associated tags) described above define an implicit anatomical and physiological taxonomy of an animal or human body whose granularity in terms of individual structures may vary depending on the application. For example, while single cells could be considered as individual structures within the taxonomy of the tagging language, given the huge number of cells in a body, exceedingly large computational resources would be required to express the body structure at such a fine level of detail. Conversely at the other end of the taxonomy, body structures can be simplified to individual systems, such as where the entire urinary system or the respiratory system can be considered as a single discrete structure.
  • Each individual structure can be represented as image data stored in a machine readable storage medium. The image data can be in any suitable format without departing from the spirit of the invention.
  • The degree of image detail for each individual structure can vary depending on the intended application. For example, the image data for a structure may be as simple as including a two-dimensional image of the structure, such as an image extracted from an X-ray scan. In another example, the image data can include a three-dimensional image of the structure, such that during visualization the image can be manipulated so that it can be seen from different perspectives.
  • Another possibility is to provide a structure that can be represented by a three-dimensional modeling program on the basis of a three-dimensional mesh. The mesh can be resized, stretched or otherwise modified to change the shape of the basic organ. The three-dimensional modeler also can include a texture-mapping feature that can apply textures onto the mesh. The three-dimensional modeler can be used to generate a three dimensional image of the outside of the structure but also can be used to generate a complete three dimensional representation of the entire structure, showing its outside surface and also its internal features as well. In the case of a human heart, for example, this form of representation could be used to show the internal structure of the human heart, therefore allowing a user to see the outside of the heart, manipulate the heart to see it from different angles, take virtual ‘slices’ (cross-sections) of the heart to expose the inside structure at a certain point or ‘fly through’ the heart in order to review its external or internal structure.
  • Yet another possibility is to provide image data that actually contains several different representations of the organ, which may be two-dimensional, three-dimensional or could be represented by a three-dimensional modeling program. In this instance, the various representations of the organ could be individually analyzed and then combined to form a single organ based on observed overlaps between the different representations or prior knowledge of the structure of the organ.
  • Each structure is further associated with a tag that contains instructions about the manner in which the image data behaves. Examples of such instructions include:
      • 1. Image modifiers that alter the image data to produce altered image data. The alterations can be dimensional alternations where the dimensions of the organ are changed and/or textural alterations where the texture of the external surface of the structure is changed. The alternations can also add or subtract components from the structure. These image modifiers can be used alone or in combination to alter the image data such as to adapt the image data to a particular subject, in other words adapt the image of the structure such that it matches the corresponding structure in the body of the subject.
      • 2. Relationship with other structures. The relationship instructions can include structural relationships allowing locating the structure properly in relation to an adjacent structure in the body. For example, when the structure is a bone, the tag may contain location instructions to specify where that bone is located with relation to other bones. In this fashion, the entire set of bones can be displayed to a user where each bone is correctly located. The relationship can also include functional relationships definitions, allowing specifying the functional group to which the structure belongs. There may be instances where the three-dimensional position of one structure with relation to another is unimportant. Rather, it is important to functionally relate a group of structures. One example is the digestive system. A functional connection exits between the mouth and the intestine as they are both components of the digestive system while they are only loosely related in terms of physical position.
      • 3. Kinetic definitions. These are instructions or parameters that define motion of structures. A kinetic definition allows animating or showing movement of the body. The motion can be as simple as the movement of a limb (e.g., motion at the elbow) or as complex as animation of a beating heart or blood flowing through veins or arteries. In the case of a simple motion, the kinetic definition specifies the mechanical parameters to define the movement, such as the structures involved, the location of the pivot point and the allowed range of motion. When more complex animation is necessary, the kinetic parameters may define fluid dynamic models to simulate blood flows through veins and arteries.
  • In order to personalize the avatar 130, the information within the HACR database 410 may be subjected one or more rules in a set of personalization rules 420. The rules 420 define certain conditions or settings that adjust the appearance or internal physiology of the avatar 130 in concordance with that observed in the corresponding subject.
  • In order to personalize the avatar 130, the information within the HACR database 410 may be subjected one or more rules in a set of personalization rules 420. The rules 420 determine how the generic avatar will be altered to match the subject. The personalization rules include logic that alters the image data associated with the respective body structures. That logic is embedded in the tags of the respective structures such that the behavior of the image data corresponding to the structures changes as desired.
  • The image alterations during a personalization process of the generic avatar are designed to perform the following, including among others, aging, changes to corporeal traits, changes based on gender or race, as well as possible exceptions. Further information about these alterations defined by the personalization rules are provided below.
  • Aging (or age adjustment) rules refer to adjustment rules that are intended to adjust the visual appearance of the set of structures comprising the avatar 130 so that they match the age of the subject.
  • In one possible form of implementation, a set of age adjustment rules exist, where different aging rules apply to different structures, as different structures are affected in a different way as a result of aging. Each age adjustment rule models the effect of aging on a structure and in particular on how a structure appears.
  • The model, which as indicated earlier, may be specific to an individual structure or may affect a set of structures, can be based on empirical observation on the effect of aging on body structures. For example, in the case of human bones, aging can affect the bone dimensions and its density. As a person ages, his or her bones are likely to shrink slightly and also become more porous.
  • To model this effect, an aging rule will typically include logic that changes the image data such that the image of the bone is resized as a function of age. As a result, the older the subject, the smaller his or her bones will appear. The degree of re-sizing can be derived from medical knowledge and observation and would generally be known to those skilled in the art.
  • Because a similar relationship is known to exist between bone density and age, another age adjustment rule for human bones may be used to depict changes to bone porosity with age. In this case, pores are created in the image (either at random positions or at predetermined positions), where the number of pores and their size is dependent on the age of the subject. As a result, the older the subject, the higher the number of pores and the larger their size will be.
  • Another example of an aging rule may relate to the pigmentation, color and texture of the skin. The age adjustment rule associated with these body structures define a texture and color model that is age-dependent. For example, the texture and color gradations can be based on empirical observations that mimic how the skin ages. As the subject gets older, the texture and color models will render the image of these structures in a way that will realistically mimic the skin of an older person on the avatar 130. For instance, the model may control the rendering of the surface of the skin, such that the skin looks rougher and may have small dark dots randomly distributed.
  • Yet another example of an age adjustment rule could be a rule that affects the appearance of the prostate gland. As is generally well known, the size of the prostate often becomes enlarged with age. The age adjustment rule would therefore be designed to alter the size (and possibly shape) of the prostate such that it becomes larger with age.
  • Another possible example of an aging rule may be one associated with gums. It is well known that as a person ages, his or her gums recede. Accordingly, the model implemented by the age adjustment rule would be designed to alter the image data of the gums such that the gums appear as receding, where the amount of receding is dependent on age.
  • In addition to changing the way the image of a structure appears to an observer, age adjustment rules can also be provided that alter certain kinetic functions which are known to be age-dependent. For instance, age typically affects the range of motion at a joint, such as the knee or elbow. To model these effects, an aging rule may be implemented that when the avatar 130 displays movement at those joints, the motion is restricted to a range that is age dependent. As a result, the avatars of higher aged subject would have a lower range of motion for the affected joints and related structures.
  • It will be appreciated that simulation of other motions can be conditioned in a similar way. For instance the general heart beat rate for the avatar 130 may be lowered as age increases to reflect known medical knowledge about the relationship between a person's heart rate and his or her age.
  • In addition to the age adjustment rules discussed above, the personalization rules engine may also include the following:
      • Corporeal Trait rules: rules that define changes to the avatar 130 based on certain corporeal traits, such as the length of arms/legs relative to the torso;
      • Gender rules: rules that define changes to the avatar 130 based on the selected genders, such as the relative location of reproductive organs and/or breast muscles/mammary glands;
      • Racial Trait rules: rules that define changes to the avatar 130 based on a selected race (where applicable or allowed), such as an adjustment of the epicanthic fold of the eyelid for those of Asian descent; and
      • Exceptions: exceptions to one or more of the above rules, which is likely based on observation or existing medical records, such as a missing arm or leg.
  • Those skilled in the art will appreciate that the above list of categories for the set of personalization rules 420 is not exclusive and that other categories and/or rules may fall within the scope of the invention.
  • FIG. 6 is a yet more detailed block diagram of the avatar evolution engine 320. The avatar evolution engine 320 operates on the basis of a set of ‘evolution’ rules that condition a set of input data to update an avatar from a prior state to a new state. The starting point of the updating process is the personalized avatar. The personalized avatar therefore is altered progressively by the updating engine such as the avatar continues to represent the subject as the body of the subject evolves over time and changes due to aging and medical conditions. Generally, the changes to the avatar made by the updating rules engine can include progressive changes such as those due to aging and discrete changes resulting from specific medical conditions encountered.
  • The set of ‘evolution’ rules that condition the input data in order to update an avatar from a prior state to a new state are represented in FIG. 6 by an updating rules engine 620. FIG. 7 shows various categories of rules that may be included within the set of evolution rules represented by the engine 620, which could include among others:
      • Aging rules: rules that define changes to the avatar 130 between states as the body of the corresponding subject ages, such as changes to the skin texture of a person as they age. These rules can be the same or similar to the aging rules discussed earlier in connection with the personalization rules;
      • Genetic rules: rules that model progressive changes to the different structures of the avatar 130 between states according to the genetic profile and makeup of the corresponding subject;
      • Demographic group rules: rules that model progressive changes to the avatar between states according to the general demographic group to which the corresponding subject belongs, such as changes known to afflict 40-45 year old white male smokers who consume between one and two packs of cigarettes per day;
      • Geographic group rules: rules that model progressive changes to the avatar between states according to the general geographic locale to which the corresponding subject belongs, such as changes due to living in a urban environment where exposure to fine particulates and other pollutants is higher than in a rural environment; and/or
      • Observed medical condition rules: rules that are generated from observed medical conditions, such as medical conditions observed from X-rays or blood tests (e.g., blood clots in the case of stroke) and generally medical observations about the medical condition of the subject.
  • It is worth noting that one or more of the rules (and in particular, the observed medical condition rules) in the updating rules engine 620 described above may originate from the medical information database 110. For example, the genetic rules may originate from the genetic profile of the corresponding subject, which may be stored in the database 110.
  • In certain cases, the aging rules in the updating rules engine 620 may be updated or altered based on contributions from the other rule categories. For example, the geographic and/or demographic group categories may cause an adjustment in the aging rules that causes the avatar 130 to age faster or slower than would otherwise be expected. For example, the aging rules for a Chinese male laborer who lives in or around central Shanghai, China and smokes at least two (2) packs of cigarettes a day would likely cause this subject's avatar to age more quickly than otherwise.
  • In contrast, the identification of certain genetic conditions in the genetic profile of a subject that confer improved resistance to certain diseases that are more common to a particular demographic group (e.g., resistance to heart disease in people 50 and above) that may be expressed in the genetic rules may cause the avatar 130 to age more slowly than would otherwise be expected.
  • The various rules within the updating rule engine 620 only govern the evolution of the avatar 130 between two states separated by time, namely a first, earlier state and a second, later state. It will be appreciated that such changes may or may not relate to the actual physiological evolution of the avatar's corresponding subject. In cases where the evolved state of the avatar 130 differs from that of its corresponding subject, the avatar 130 may be further updated based on observed medical conditions.
  • FIG. 8 illustrates a non-limiting method by which the updating rule engine 620 may update the avatar 130 between a first prior state and a second, more current state based on observed medical conditions. This figure includes two (2) sets of data, namely a medical non-image dataset 810 and a medical image dataset 820. Although the datasets 810 and 820 are presented here as separate entities, this is done for the sake of illustration. In reality, both of these datasets are quite likely to reside together in the medical information database 110.
  • The contents of the medical non-image dataset 810 typically contain medical information for the subject that is non-visual in nature, such as numeric test results, observation notes by medical personnel and/or biopsy reports, among others. Moreover, contents of this dataset may be linked to certain aspects of the HACR database 410, such as tagged content within the structures component 412 and/or internal kinetics component 414. For example, a test showing the blood pressure and flow through specific arteries in the cardiovascular system may be used to model blood flow in the avatar 130.
  • In contrast, the contents of the medical image dataset 820 include medical information that is visual in nature, such as X-ray images, photographs taken from a biopsy and/or CT-scan related data and ultrasound observations among others. Furthermore, contents of this dataset may be linked to or associated with the HACR database 410 in order to provide the images used for tagged structures within the structures component 412. For example, an X-ray of a leg bone and surrounding tissue may be associated with the tagged structure that defines how the leg bone and surrounding tissue in the avatar 130 is represented.
  • The avatar evolution engine 320 may monitor the contents of the datasets 810 and 820 in order that it can become aware of any new information that is added to these datasets from observation of the subject. Alternatively, the engine 320 may be advised of the addition of new data to the datasets 810 and 820 only at the time when the avatar 130 is to be updated.
  • Once the avatar evolution engine 320 becomes aware of new information in the datasets 810 and 820, it can use this information to update the observed medical condition rules components of the updating rules engine 620 in order to update the avatar 130 in a similar fashion.
  • In particular, new information within the medical non-image dataset 810 could be used to update a set of non-image based updating rules 815 that may be included within the observed medical condition rules category in the updating rules engine 620. Similarly, new information within the medical image dataset 820 could be used to update a set of image-based updating rules 825, which may also be included within the observed medical condition rules category in the updating rules engine 620.
  • For example, assume that a subject suffers a fall and that their brain is subjected to a CT scan to ensure that they are not suffering from a condition, such as a concussion or brain edema. The data generated by the CT scan of the subject's brain is stored within the medical information database 110 and becomes part of the medical image dataset 820 as a result.
  • The addition of this data to the dataset 820 may trigger the avatar evolution engine 320 to review and revise the updating rules engine 620 based on this new information. In particular, the engine 620 may use this data to update the observed medical condition rules category for the brain to update the previous image associated with the tagged brain entry in the structures component 410 (which would likely have been taken before the fall) with a new image updated to take into account the information contained in the CT scan data. Because the avatar evolution engine 320 now has two separate brain images from the subject, it can evaluate changes between the images in order to update the brain represented in the avatar 130 in the same way. This can be done by updating the observed medical condition category of the updating rules engine 620 to account for the new image, which may involve adjusting the image modifier information for the tag associated with the brain structure.
  • Although the above example used new image data within the medical image dataset 820 as the trigger for the update of the updating rules engine 620 by the avatar evolution engine 320, those skilled in the art will understand that a similar process could be used to update the non-image updating rules 815 based on information added to the medical non-image dataset 810.
  • FIG. 9 shows a flowchart that illustrates a non-limiting process by which information in the previously mentioned medical image dataset 820 (and/or the medical information database 110 as a whole) could be used to update the avatar 130.
  • At step 910, image data is processed to identify the particular structure (i.e., body part or organ of the avatar) to which the image applies. This data may be processed by the avatar evolution engine 320, by the updating rules engine 620 or by an image processing software module that is likely similar (if not identical) to discussed in the context of avatar personalization.
  • In certain cases, the structure or body part to which the image applies may be included within the image data. For example, an image taken of a femur bone may include metadata (which may be based on BML) that may indicate the image was of a left femur bone. Among other image-related information that might be provided within the image data may include the angle at which the image was taken, the device used to generate the image and/or an indication as to why the image was generated (e.g., as part of a standard checkup or as a result of certain trauma).
  • If such information is included within the image data, the process may proceed to the next step immediately. However, if this information is missing from or is not included with the image data, it may be extracted at this point by analyzing the medical image and comparing any features identified within it against known body structures and/or structures within the subject's body in particular.
  • For example, if a bone is identified within the image (such as by comparing adjacent gray-level gradient values), the size and shape of the bone may be compared to those found in existing images and/or models to see which of these produce the closest match. Returning briefly to the example of the femur X-ray mentioned above, if the image data for the X-ray did not include information defining the imaged bone as a femur, the identified bone may be compared against bones within the avatar 130 and/or medical image dataset 820, and more specifically, images that contain bones with a similar size, shape and/or orientation.
  • Since it is believed that knowledge of image processing and pattern matching techniques to achieve this result are known in the art, a further description of how this matching occurs will not be provided here. However, it is worth noting that the image processing and/or pattern matching may be performed against bones associated with the avatar 130 of the subject, against bones associated with the medical image dataset 820 or against bones known to be in images stored within the medical information database 110. This can increase the likelihood that a bone that is captured within an image will be matched correctly to its corresponding structure.
  • At step 920, relevant features are extracted from the image data. During this step, the image is processed to identify relevant features in the structure, which may include among others:
      • breaks or separations in the structure, such as from a broken bone;
      • changes in the dimensions, shape and/or density, such as those due to age;
      • unexpected growths or abscesses that might indicate disease, such as cancerous growths or tumours;
  • The process by which relevant features may be identified may include comparing the structure within in the current image with an image of the structure taken at a prior state, which may be stored within the medical image dataset 820. For example, an X-ray image of a subject's femur may be compared against earlier X-ray images of the same bone to identify any changes that have taken place.
  • It is worth noting that although steps 910 and 920 in FIG. 9 are shown in sequential order, the processing of the image that occurs in these steps may also be performed more or less simultaneously. Therefore, while the image is being processed to identify its corresponding structure (step 910), it may also be processed simultaneously to identify relevant features (step 920).
  • The result of the previous step was the identification of relevant features for the structure based on image data from the subject. In order to ensure the avatar 130 reflects the state of its corresponding subject, the avatar must be updated in the same way.
  • At step 930, the avatar 130 is updated to include the same relevant features as were identified during the previous step. This update to the avatar 130 is typically done by the avatar evolution engine 320 via the updating rules engine 620. More specifically, the update may be performed by the engine 320 using the updated non-image based input rules 815 and image-based input rules 825 of the observed medical conditions rule category residing within the engine 620.
  • Upon the completion of step 930, the avatar 130 will have been updated to reflect the most current medical condition of its corresponding subject. This process prepares the avatar 130 for viewing by medical personnel in order to diagnose and/or treat medical conditions affecting the subject.
  • FIG. 11 is a block diagram of an image viewer that can used for viewing the updated avatar in its entirety of components thereof. Generally, the image viewer is associated with the user interface of the medical record and a user can invoke the viewer from the user interface control. The viewer includes a body structures module 1020 that allows the user to select the particular structure or set of structures for display. For instance, the user can select a single structure to be shown, such as a bone or an organ, say the heart. The viewer can provide navigational tools allowing the user to rotate the image such that it can be seen from different perspectives, create slices to see the inside of the structure, among others. In the specific example shown, a slice through the entire body of the avatar is illustrated.
  • In addition to showing individual structures, the viewer allows the user to display a series of structures that are related to one another, either by virtue of physical relation or functional relation.
  • The viewer module also has a kinetics viewer that can show animation of selected structures. For instance the kinetics viewer can animate a joint and depict how the various bones move, simulate the beating of the heart, simulate the blood flow through a certain organ, etc.
  • Although various embodiments have been illustrated, this was for the purpose of describing, but not limiting, the invention. Various modifications will become apparent to those skilled in the art and are within the scope of this invention, which is defined more particularly by the attached claims.

Claims (1)

1. A medical records system, comprising:
a. a data processing apparatus including a CPU and a computer readable storage medium encoded with a program for execution by the CPU, the program processing medical information in connection with a subject to generate an avatar of the subject which conveys the medical information;
b. the data processing apparatus having an output for releasing data representative of the avatar.
US12/726,959 2009-03-18 2010-03-18 Medical records system with dynamic avatar generator and avatar viewer Abandoned US20100257214A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/726,959 US20100257214A1 (en) 2009-03-18 2010-03-18 Medical records system with dynamic avatar generator and avatar viewer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16129909P 2009-03-18 2009-03-18
US12/726,959 US20100257214A1 (en) 2009-03-18 2010-03-18 Medical records system with dynamic avatar generator and avatar viewer

Publications (1)

Publication Number Publication Date
US20100257214A1 true US20100257214A1 (en) 2010-10-07

Family

ID=42735931

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/726,959 Abandoned US20100257214A1 (en) 2009-03-18 2010-03-18 Medical records system with dynamic avatar generator and avatar viewer

Country Status (2)

Country Link
US (1) US20100257214A1 (en)
CA (1) CA2697309A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120329561A1 (en) * 2010-12-09 2012-12-27 Genomic Arts, LLC System and methods for generating avatars and art
US20130069940A1 (en) * 2011-09-21 2013-03-21 University Of South Florida (A Florida Non-Profit Corporation) Systems And Methods For Projecting Images Onto An Object
US20130315452A1 (en) * 2012-05-22 2013-11-28 Orca Health, Inc. Personalized anatomical diagnostics and simulations
US8972882B2 (en) 2013-01-30 2015-03-03 Orca Health, Inc. User interfaces and systems for oral hygiene
US8992232B2 (en) 2011-09-20 2015-03-31 Orca Health, Inc. Interactive and educational vision interfaces
US9256962B2 (en) 2013-01-23 2016-02-09 Orca Health Inc. Personalizing medical conditions with augmented reality
US20160045171A1 (en) * 2013-03-15 2016-02-18 General Electric Company Methods and systems for improving patient engagement via medical avatars
US20170352194A1 (en) * 2016-06-06 2017-12-07 Biodigital, Inc. Methodology & system for mapping a virtual human body
WO2019165432A1 (en) * 2018-02-26 2019-08-29 Children's Medical Center Corporation Extended reality medical report generating system
US10546656B2 (en) * 2014-05-09 2020-01-28 Acupath Laboratories, Inc. Biopsy mapping tools
US11291919B2 (en) * 2017-05-07 2022-04-05 Interlake Research, Llc Development of virtual character in a learning game
US11727724B1 (en) 2018-09-27 2023-08-15 Apple Inc. Emotion detection
US11801446B2 (en) * 2019-03-15 2023-10-31 Sony Interactive Entertainment Inc. Systems and methods for training an artificial intelligence model for competition matches
US11830182B1 (en) * 2019-08-20 2023-11-28 Apple Inc. Machine learning-based blood flow tracking
US11869150B1 (en) 2017-06-01 2024-01-09 Apple Inc. Avatar modeling and generation

Citations (96)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4351982A (en) * 1980-12-15 1982-09-28 Racal-Milgo, Inc. RSA Public-key data encryption system having large random prime number generating microprocessor or the like
US4658093A (en) * 1983-07-11 1987-04-14 Hellman Martin E Software distribution system
US4704610A (en) * 1985-12-16 1987-11-03 Smith Michel R Emergency vehicle warning and traffic control system
US4796220A (en) * 1986-12-15 1989-01-03 Pride Software Development Corp. Method of controlling the copying of software
US4893270A (en) * 1986-05-12 1990-01-09 American Telephone And Telegraph Company, At&T Bell Laboratories Medical information system
US5210795A (en) * 1992-01-10 1993-05-11 Digital Equipment Corporation Secure user authentication from personal computer
US5291598A (en) * 1992-04-07 1994-03-01 Gregory Grundy Method and system for decentralized manufacture of copy-controlled software
US5414269A (en) * 1991-10-29 1995-05-09 Oki Electric Industry Co., Ltd. Circuit for detecting a paper at a desired position along a paper feed path with a one shot multivibrator actuating circuit
US5418854A (en) * 1992-04-28 1995-05-23 Digital Equipment Corporation Method and apparatus for protecting the confidentiality of passwords in a distributed data processing system
US5440635A (en) * 1993-08-23 1995-08-08 At&T Corp. Cryptographic protocol for remote authentication
US5490216A (en) * 1992-09-21 1996-02-06 Uniloc Private Limited System for software registration
US5666415A (en) * 1995-07-28 1997-09-09 Digital Equipment Corporation Method and apparatus for cryptographic authentication
US5715823A (en) * 1996-02-27 1998-02-10 Atlantis Diagnostics International, L.L.C. Ultrasonic diagnostic imaging system with universal access to diagnostic information and images
US5745879A (en) * 1991-05-08 1998-04-28 Digital Equipment Corporation Method and system for managing execution of licensed programs
US5754763A (en) * 1996-10-01 1998-05-19 International Business Machines Corporation Software auditing mechanism for a distributed computer enterprise environment
US5790664A (en) * 1996-02-26 1998-08-04 Network Engineering Software, Inc. Automated system for management of licensed software
US5845255A (en) * 1994-10-28 1998-12-01 Advanced Health Med-E-Systems Corporation Prescription management system
US5884246A (en) * 1996-12-04 1999-03-16 Transgate Intellectual Properties Ltd. System and method for transparent translation of electronically transmitted messages
US5903889A (en) * 1997-06-09 1999-05-11 Telaric, Inc. System and method for translating, collecting and archiving patient records
US5918010A (en) * 1997-02-07 1999-06-29 General Internet, Inc. Collaborative internet data mining systems
US5924074A (en) * 1996-09-27 1999-07-13 Azron Incorporated Electronic medical records system
US5925127A (en) * 1997-04-09 1999-07-20 Microsoft Corporation Method and system for monitoring the use of rented software
US5974150A (en) * 1997-09-30 1999-10-26 Tracer Detection Technology Corp. System and method for authentication of goods
US6009401A (en) * 1998-04-06 1999-12-28 Preview Systems, Inc. Relicensing of electronically purchased software
US6012083A (en) * 1996-09-24 2000-01-04 Ricoh Company Ltd. Method and apparatus for document processing using agents to process transactions created based on document content
US6018713A (en) * 1997-04-09 2000-01-25 Coli; Robert D. Integrated system and method for ordering and cumulative results reporting of medical tests
US6044471A (en) * 1998-06-04 2000-03-28 Z4 Technologies, Inc. Method and apparatus for securing software to reduce unauthorized use
US6158005A (en) * 1998-09-10 2000-12-05 Audible, Inc. Cloning protection scheme for a digital information playback device
US6230199B1 (en) * 1999-10-29 2001-05-08 Mcafee.Com, Inc. Active marketing based on client computer configurations
US6233567B1 (en) * 1997-08-29 2001-05-15 Intel Corporation Method and apparatus for software licensing electronically distributed programs
US6243468B1 (en) * 1998-04-29 2001-06-05 Microsoft Corporation Software anti-piracy system that adapts to hardware upgrades
US6263330B1 (en) * 1998-02-24 2001-07-17 Luc Bessette Method and apparatus for the management of data files
US6294793B1 (en) * 1992-12-03 2001-09-25 Brown & Sharpe Surface Inspection Systems, Inc. High speed optical inspection apparatus for a transparent disk using gaussian distribution analysis and method therefor
US20010034712A1 (en) * 1998-06-04 2001-10-25 Colvin David S. System and method for monitoring software
US20010044782A1 (en) * 1998-04-29 2001-11-22 Microsoft Corporation Hardware ID to prevent software piracy
US6330670B1 (en) * 1998-10-26 2001-12-11 Microsoft Corporation Digital rights management operating system
US20020019814A1 (en) * 2001-03-01 2002-02-14 Krishnamurthy Ganesan Specifying rights in a digital rights license according to events
US6364834B1 (en) * 1996-11-13 2002-04-02 Criticare Systems, Inc. Method and system for remotely monitoring multiple medical parameters in an integrated medical monitoring system
US20020082997A1 (en) * 2000-07-14 2002-06-27 Hiroshi Kobata Controlling and managing digital assets
US6449645B1 (en) * 1999-01-19 2002-09-10 Kenneth L. Nash System for monitoring the association of digitized information having identification indicia with more than one of uniquely identified computers in a network for illegal use detection
US20020161718A1 (en) * 1998-08-04 2002-10-31 Coley Christopher D. Automated system for management of licensed software
US6536005B1 (en) * 1999-10-26 2003-03-18 Teradyne, Inc. High-speed failure capture apparatus and method for automatic test equipment
US20030065918A1 (en) * 2001-04-06 2003-04-03 Willey William Daniel Device authentication in a PKI
US20030172035A1 (en) * 2002-03-08 2003-09-11 Cronce Paul A. Method and system for managing software licenses
US20040024860A1 (en) * 2000-10-26 2004-02-05 Katsuhiko Sato Communication system, terminal, reproduction program, recorded medium on which reproduction program is recorded, server device, server program, and recorded medium on which server program is recorded
US20040030912A1 (en) * 2001-05-09 2004-02-12 Merkle James A. Systems and methods for the prevention of unauthorized use and manipulation of digital content
US20040059929A1 (en) * 2000-09-14 2004-03-25 Alastair Rodgers Digital rights management
US20040122787A1 (en) * 2002-12-18 2004-06-24 Avinash Gopal B. Enhanced computer-assisted medical data processing system and method
US20040143746A1 (en) * 2003-01-16 2004-07-22 Jean-Alfred Ligeti Software license compliance system and method
US6775670B2 (en) * 1998-05-29 2004-08-10 Luc Bessette Method and apparatus for the management of data files
US20040187018A1 (en) * 2001-10-09 2004-09-23 Owen William N. Multi-factor authentication system
US6859793B1 (en) * 2002-12-19 2005-02-22 Networks Associates Technology, Inc. Software license reporting and control system and method
US20050108173A1 (en) * 1994-11-23 2005-05-19 Contentgurad Holdings, Inc. System for controlling the distribution and use digital works using digital tickets
US20050138155A1 (en) * 2003-12-19 2005-06-23 Michael Lewis Signal assessment
US6920567B1 (en) * 1999-04-07 2005-07-19 Viatech Technologies Inc. System and embedded license control mechanism for the creation and distribution of digital content files and enforcement of licensed use of the digital content files
US20050172280A1 (en) * 2004-01-29 2005-08-04 Ziegler Jeremy R. System and method for preintegration of updates to an operating system
US6976009B2 (en) * 2001-05-31 2005-12-13 Contentguard Holdings, Inc. Method and apparatus for assigning consequential rights to documents and documents having such rights
US20060072444A1 (en) * 2004-09-29 2006-04-06 Engel David B Marked article and method of making the same
US7028049B1 (en) * 1996-02-17 2006-04-11 Allcare Health Management System, Inc. Standing order database search system and method for internet and internet application
US7032110B1 (en) * 2000-06-30 2006-04-18 Landesk Software Limited PKI-based client/server authentication
US20060095454A1 (en) * 2004-10-29 2006-05-04 Texas Instruments Incorporated System and method for secure collaborative terminal identity authentication between a wireless communication device and a wireless operator
US20060136143A1 (en) * 2004-12-17 2006-06-22 General Electric Company Personalized genetic-based analysis of medical conditions
US7069595B2 (en) * 2001-03-23 2006-06-27 International Business Machines Corporation Method of controlling use of digitally encoded products
US7069440B2 (en) * 2000-06-09 2006-06-27 Northrop Grumman Corporation Technique for obtaining a single sign-on certificate from a foreign PKI system using an existing strong authentication PKI system
US20060161914A1 (en) * 2005-01-14 2006-07-20 Microsoft Corporation Systems and methods to modify application installations
US7085741B2 (en) * 2001-01-17 2006-08-01 Contentguard Holdings, Inc. Method and apparatus for managing digital content usage rights
US20060184489A1 (en) * 2004-12-17 2006-08-17 General Electric Company Genetic knowledgebase creation for personalized analysis of medical conditions
US20060265337A1 (en) * 1996-02-26 2006-11-23 Graphon Corporation Automated system for management of licensed digital assets
US20060282511A1 (en) * 2005-06-14 2006-12-14 Hitachi Global Storage Technologies Netherlands B.V. Method for limiting utilizing terminal of contents, and memory device and system for method
US7188241B2 (en) * 2002-10-16 2007-03-06 Pace Antipiracy Protecting software from unauthorized use by applying machine-dependent modifications to code modules
US7203966B2 (en) * 2001-06-27 2007-04-10 Microsoft Corporation Enforcement architecture and method for digital rights management system for roaming a license to a plurality of user devices
US7206765B2 (en) * 2001-01-17 2007-04-17 Contentguard Holdings, Inc. System and method for supplying and managing usage rights based on rules
US20070168288A1 (en) * 2006-01-13 2007-07-19 Trails.Com, Inc. Method and system for dynamic digital rights bundling
US20070198422A1 (en) * 2005-12-19 2007-08-23 Anand Prahlad System and method for providing a flexible licensing system for digital content
US7272728B2 (en) * 2004-06-14 2007-09-18 Iovation, Inc. Network security and fraud detection system and method
US20070219917A1 (en) * 2004-03-29 2007-09-20 Smart Internet Tecnoogy Crc Pty Limited Digital License Sharing System and Method
US20070282615A1 (en) * 2006-06-01 2007-12-06 Hamilton Rick A Method for Digital Rights Management
US7319987B1 (en) * 1996-08-29 2008-01-15 Indivos Corporation Tokenless financial access system
US7327280B2 (en) * 2002-08-15 2008-02-05 California Institute Of Technology Emergency vehicle traffic signal preemption system
US7337147B2 (en) * 2005-06-30 2008-02-26 Microsoft Corporation Dynamic digital content licensing
US7343297B2 (en) * 2001-06-15 2008-03-11 Microsoft Corporation System and related methods for managing and enforcing software licenses
US20080065552A1 (en) * 2006-09-13 2008-03-13 Gidon Elazar Marketplace for Transferring Licensed Digital Content
US20080086423A1 (en) * 2006-10-06 2008-04-10 Nigel Waites Media player with license expiration warning
US20080147556A1 (en) * 2006-12-15 2008-06-19 Nbc Universal, Inc. Digital rights management flexible continued usage system and method
US20080228578A1 (en) * 2007-01-25 2008-09-18 Governing Dynamics, Llc Digital rights management and data license management
US7463945B2 (en) * 2001-07-13 2008-12-09 Siemens Aktiengesellschaft Electronic fingerprints for machine control and production machines
US20080320607A1 (en) * 2007-06-21 2008-12-25 Uniloc Usa System and method for auditing software usage
US20090028380A1 (en) * 2007-07-23 2009-01-29 Hillebrand Greg Method and apparatus for realistic simulation of wrinkle aging and de-aging
US20090083730A1 (en) * 2007-09-20 2009-03-26 Richardson Ric B Installing Protected Software Product Using Unprotected Installation Image
US20090138975A1 (en) * 2007-11-17 2009-05-28 Uniloc Usa System and Method for Adjustable Licensing of Digital Products
US20090164917A1 (en) * 2007-12-19 2009-06-25 Kelly Kevin M System and method for remote delivery of healthcare and treatment services
US7653899B1 (en) * 2004-07-23 2010-01-26 Green Hills Software, Inc. Post-execution software debugger with performance display
US20100128946A1 (en) * 2008-11-22 2010-05-27 General Electric Company Systems, apparatus and processes for automated medical image segmentation using a statistical model
US7734656B2 (en) * 1998-02-24 2010-06-08 Luc Bessette System and method for electronically managing medical data files in order to facilitate genetic research
US20100217619A1 (en) * 2009-02-26 2010-08-26 Aaron Roger Cox Methods for virtual world medical symptom identification
US20120004894A1 (en) * 2007-09-21 2012-01-05 Edwin Brian Butler Systems, Methods and Apparatuses for Generating and using Representations of Individual or Aggregate Human Medical Data

Patent Citations (98)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4351982A (en) * 1980-12-15 1982-09-28 Racal-Milgo, Inc. RSA Public-key data encryption system having large random prime number generating microprocessor or the like
US4658093A (en) * 1983-07-11 1987-04-14 Hellman Martin E Software distribution system
US4704610A (en) * 1985-12-16 1987-11-03 Smith Michel R Emergency vehicle warning and traffic control system
US4893270A (en) * 1986-05-12 1990-01-09 American Telephone And Telegraph Company, At&T Bell Laboratories Medical information system
US4796220A (en) * 1986-12-15 1989-01-03 Pride Software Development Corp. Method of controlling the copying of software
US5745879A (en) * 1991-05-08 1998-04-28 Digital Equipment Corporation Method and system for managing execution of licensed programs
US5414269A (en) * 1991-10-29 1995-05-09 Oki Electric Industry Co., Ltd. Circuit for detecting a paper at a desired position along a paper feed path with a one shot multivibrator actuating circuit
US5210795A (en) * 1992-01-10 1993-05-11 Digital Equipment Corporation Secure user authentication from personal computer
US5291598A (en) * 1992-04-07 1994-03-01 Gregory Grundy Method and system for decentralized manufacture of copy-controlled software
US5418854A (en) * 1992-04-28 1995-05-23 Digital Equipment Corporation Method and apparatus for protecting the confidentiality of passwords in a distributed data processing system
US5490216A (en) * 1992-09-21 1996-02-06 Uniloc Private Limited System for software registration
US6294793B1 (en) * 1992-12-03 2001-09-25 Brown & Sharpe Surface Inspection Systems, Inc. High speed optical inspection apparatus for a transparent disk using gaussian distribution analysis and method therefor
US5440635A (en) * 1993-08-23 1995-08-08 At&T Corp. Cryptographic protocol for remote authentication
US5845255A (en) * 1994-10-28 1998-12-01 Advanced Health Med-E-Systems Corporation Prescription management system
US20050108173A1 (en) * 1994-11-23 2005-05-19 Contentgurad Holdings, Inc. System for controlling the distribution and use digital works using digital tickets
US5666415A (en) * 1995-07-28 1997-09-09 Digital Equipment Corporation Method and apparatus for cryptographic authentication
US7028049B1 (en) * 1996-02-17 2006-04-11 Allcare Health Management System, Inc. Standing order database search system and method for internet and internet application
US5790664A (en) * 1996-02-26 1998-08-04 Network Engineering Software, Inc. Automated system for management of licensed software
US20060265337A1 (en) * 1996-02-26 2006-11-23 Graphon Corporation Automated system for management of licensed digital assets
US5715823A (en) * 1996-02-27 1998-02-10 Atlantis Diagnostics International, L.L.C. Ultrasonic diagnostic imaging system with universal access to diagnostic information and images
US7319987B1 (en) * 1996-08-29 2008-01-15 Indivos Corporation Tokenless financial access system
US6012083A (en) * 1996-09-24 2000-01-04 Ricoh Company Ltd. Method and apparatus for document processing using agents to process transactions created based on document content
US5924074A (en) * 1996-09-27 1999-07-13 Azron Incorporated Electronic medical records system
US5754763A (en) * 1996-10-01 1998-05-19 International Business Machines Corporation Software auditing mechanism for a distributed computer enterprise environment
US6364834B1 (en) * 1996-11-13 2002-04-02 Criticare Systems, Inc. Method and system for remotely monitoring multiple medical parameters in an integrated medical monitoring system
US5884246A (en) * 1996-12-04 1999-03-16 Transgate Intellectual Properties Ltd. System and method for transparent translation of electronically transmitted messages
US5918010A (en) * 1997-02-07 1999-06-29 General Internet, Inc. Collaborative internet data mining systems
US5925127A (en) * 1997-04-09 1999-07-20 Microsoft Corporation Method and system for monitoring the use of rented software
US6018713A (en) * 1997-04-09 2000-01-25 Coli; Robert D. Integrated system and method for ordering and cumulative results reporting of medical tests
US5903889A (en) * 1997-06-09 1999-05-11 Telaric, Inc. System and method for translating, collecting and archiving patient records
US6233567B1 (en) * 1997-08-29 2001-05-15 Intel Corporation Method and apparatus for software licensing electronically distributed programs
US5974150A (en) * 1997-09-30 1999-10-26 Tracer Detection Technology Corp. System and method for authentication of goods
US6263330B1 (en) * 1998-02-24 2001-07-17 Luc Bessette Method and apparatus for the management of data files
US7734656B2 (en) * 1998-02-24 2010-06-08 Luc Bessette System and method for electronically managing medical data files in order to facilitate genetic research
US6009401A (en) * 1998-04-06 1999-12-28 Preview Systems, Inc. Relicensing of electronically purchased software
US20010044782A1 (en) * 1998-04-29 2001-11-22 Microsoft Corporation Hardware ID to prevent software piracy
US6243468B1 (en) * 1998-04-29 2001-06-05 Microsoft Corporation Software anti-piracy system that adapts to hardware upgrades
US6775670B2 (en) * 1998-05-29 2004-08-10 Luc Bessette Method and apparatus for the management of data files
US6044471A (en) * 1998-06-04 2000-03-28 Z4 Technologies, Inc. Method and apparatus for securing software to reduce unauthorized use
US6785825B2 (en) * 1998-06-04 2004-08-31 Z4 Technologies, Inc. Method for securing software to decrease software piracy
US20010034712A1 (en) * 1998-06-04 2001-10-25 Colvin David S. System and method for monitoring software
US20020161718A1 (en) * 1998-08-04 2002-10-31 Coley Christopher D. Automated system for management of licensed software
US6158005A (en) * 1998-09-10 2000-12-05 Audible, Inc. Cloning protection scheme for a digital information playback device
US6330670B1 (en) * 1998-10-26 2001-12-11 Microsoft Corporation Digital rights management operating system
US6449645B1 (en) * 1999-01-19 2002-09-10 Kenneth L. Nash System for monitoring the association of digitized information having identification indicia with more than one of uniquely identified computers in a network for illegal use detection
US6920567B1 (en) * 1999-04-07 2005-07-19 Viatech Technologies Inc. System and embedded license control mechanism for the creation and distribution of digital content files and enforcement of licensed use of the digital content files
US6536005B1 (en) * 1999-10-26 2003-03-18 Teradyne, Inc. High-speed failure capture apparatus and method for automatic test equipment
US6230199B1 (en) * 1999-10-29 2001-05-08 Mcafee.Com, Inc. Active marketing based on client computer configurations
US7069440B2 (en) * 2000-06-09 2006-06-27 Northrop Grumman Corporation Technique for obtaining a single sign-on certificate from a foreign PKI system using an existing strong authentication PKI system
US7032110B1 (en) * 2000-06-30 2006-04-18 Landesk Software Limited PKI-based client/server authentication
US20020082997A1 (en) * 2000-07-14 2002-06-27 Hiroshi Kobata Controlling and managing digital assets
US20040059929A1 (en) * 2000-09-14 2004-03-25 Alastair Rodgers Digital rights management
US20040024860A1 (en) * 2000-10-26 2004-02-05 Katsuhiko Sato Communication system, terminal, reproduction program, recorded medium on which reproduction program is recorded, server device, server program, and recorded medium on which server program is recorded
US7206765B2 (en) * 2001-01-17 2007-04-17 Contentguard Holdings, Inc. System and method for supplying and managing usage rights based on rules
US7085741B2 (en) * 2001-01-17 2006-08-01 Contentguard Holdings, Inc. Method and apparatus for managing digital content usage rights
US20020019814A1 (en) * 2001-03-01 2002-02-14 Krishnamurthy Ganesan Specifying rights in a digital rights license according to events
US7069595B2 (en) * 2001-03-23 2006-06-27 International Business Machines Corporation Method of controlling use of digitally encoded products
US20030065918A1 (en) * 2001-04-06 2003-04-03 Willey William Daniel Device authentication in a PKI
US20040030912A1 (en) * 2001-05-09 2004-02-12 Merkle James A. Systems and methods for the prevention of unauthorized use and manipulation of digital content
US6976009B2 (en) * 2001-05-31 2005-12-13 Contentguard Holdings, Inc. Method and apparatus for assigning consequential rights to documents and documents having such rights
US7343297B2 (en) * 2001-06-15 2008-03-11 Microsoft Corporation System and related methods for managing and enforcing software licenses
US7203966B2 (en) * 2001-06-27 2007-04-10 Microsoft Corporation Enforcement architecture and method for digital rights management system for roaming a license to a plurality of user devices
US7463945B2 (en) * 2001-07-13 2008-12-09 Siemens Aktiengesellschaft Electronic fingerprints for machine control and production machines
US20040187018A1 (en) * 2001-10-09 2004-09-23 Owen William N. Multi-factor authentication system
US20030172035A1 (en) * 2002-03-08 2003-09-11 Cronce Paul A. Method and system for managing software licenses
US7327280B2 (en) * 2002-08-15 2008-02-05 California Institute Of Technology Emergency vehicle traffic signal preemption system
US7188241B2 (en) * 2002-10-16 2007-03-06 Pace Antipiracy Protecting software from unauthorized use by applying machine-dependent modifications to code modules
US20040122787A1 (en) * 2002-12-18 2004-06-24 Avinash Gopal B. Enhanced computer-assisted medical data processing system and method
US6859793B1 (en) * 2002-12-19 2005-02-22 Networks Associates Technology, Inc. Software license reporting and control system and method
US20040143746A1 (en) * 2003-01-16 2004-07-22 Jean-Alfred Ligeti Software license compliance system and method
US20050138155A1 (en) * 2003-12-19 2005-06-23 Michael Lewis Signal assessment
US20050172280A1 (en) * 2004-01-29 2005-08-04 Ziegler Jeremy R. System and method for preintegration of updates to an operating system
US20070219917A1 (en) * 2004-03-29 2007-09-20 Smart Internet Tecnoogy Crc Pty Limited Digital License Sharing System and Method
US7272728B2 (en) * 2004-06-14 2007-09-18 Iovation, Inc. Network security and fraud detection system and method
US7653899B1 (en) * 2004-07-23 2010-01-26 Green Hills Software, Inc. Post-execution software debugger with performance display
US20060072444A1 (en) * 2004-09-29 2006-04-06 Engel David B Marked article and method of making the same
US20060095454A1 (en) * 2004-10-29 2006-05-04 Texas Instruments Incorporated System and method for secure collaborative terminal identity authentication between a wireless communication device and a wireless operator
US20060136143A1 (en) * 2004-12-17 2006-06-22 General Electric Company Personalized genetic-based analysis of medical conditions
US20060184489A1 (en) * 2004-12-17 2006-08-17 General Electric Company Genetic knowledgebase creation for personalized analysis of medical conditions
US20060161914A1 (en) * 2005-01-14 2006-07-20 Microsoft Corporation Systems and methods to modify application installations
US20060282511A1 (en) * 2005-06-14 2006-12-14 Hitachi Global Storage Technologies Netherlands B.V. Method for limiting utilizing terminal of contents, and memory device and system for method
US7337147B2 (en) * 2005-06-30 2008-02-26 Microsoft Corporation Dynamic digital content licensing
US20070203846A1 (en) * 2005-12-19 2007-08-30 Srinivas Kavuri System and method for providing a flexible licensing system for digital content
US20070198422A1 (en) * 2005-12-19 2007-08-23 Anand Prahlad System and method for providing a flexible licensing system for digital content
US20070168288A1 (en) * 2006-01-13 2007-07-19 Trails.Com, Inc. Method and system for dynamic digital rights bundling
US20070282615A1 (en) * 2006-06-01 2007-12-06 Hamilton Rick A Method for Digital Rights Management
US20080065552A1 (en) * 2006-09-13 2008-03-13 Gidon Elazar Marketplace for Transferring Licensed Digital Content
US20080086423A1 (en) * 2006-10-06 2008-04-10 Nigel Waites Media player with license expiration warning
US20080147556A1 (en) * 2006-12-15 2008-06-19 Nbc Universal, Inc. Digital rights management flexible continued usage system and method
US20080228578A1 (en) * 2007-01-25 2008-09-18 Governing Dynamics, Llc Digital rights management and data license management
US20080320607A1 (en) * 2007-06-21 2008-12-25 Uniloc Usa System and method for auditing software usage
US20090028380A1 (en) * 2007-07-23 2009-01-29 Hillebrand Greg Method and apparatus for realistic simulation of wrinkle aging and de-aging
US20090083730A1 (en) * 2007-09-20 2009-03-26 Richardson Ric B Installing Protected Software Product Using Unprotected Installation Image
US20120004894A1 (en) * 2007-09-21 2012-01-05 Edwin Brian Butler Systems, Methods and Apparatuses for Generating and using Representations of Individual or Aggregate Human Medical Data
US20090138975A1 (en) * 2007-11-17 2009-05-28 Uniloc Usa System and Method for Adjustable Licensing of Digital Products
US20090164917A1 (en) * 2007-12-19 2009-06-25 Kelly Kevin M System and method for remote delivery of healthcare and treatment services
US20100128946A1 (en) * 2008-11-22 2010-05-27 General Electric Company Systems, apparatus and processes for automated medical image segmentation using a statistical model
US20100217619A1 (en) * 2009-02-26 2010-08-26 Aaron Roger Cox Methods for virtual world medical symptom identification

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120329561A1 (en) * 2010-12-09 2012-12-27 Genomic Arts, LLC System and methods for generating avatars and art
US8992232B2 (en) 2011-09-20 2015-03-31 Orca Health, Inc. Interactive and educational vision interfaces
US9520072B2 (en) * 2011-09-21 2016-12-13 University Of South Florida Systems and methods for projecting images onto an object
US20130069940A1 (en) * 2011-09-21 2013-03-21 University Of South Florida (A Florida Non-Profit Corporation) Systems And Methods For Projecting Images Onto An Object
US20130315452A1 (en) * 2012-05-22 2013-11-28 Orca Health, Inc. Personalized anatomical diagnostics and simulations
US8908943B2 (en) * 2012-05-22 2014-12-09 Orca Health, Inc. Personalized anatomical diagnostics and simulations
US9256962B2 (en) 2013-01-23 2016-02-09 Orca Health Inc. Personalizing medical conditions with augmented reality
US9715753B2 (en) 2013-01-23 2017-07-25 Orca Health, Inc. Personalizing medical conditions with augmented reality
US8972882B2 (en) 2013-01-30 2015-03-03 Orca Health, Inc. User interfaces and systems for oral hygiene
US20160045171A1 (en) * 2013-03-15 2016-02-18 General Electric Company Methods and systems for improving patient engagement via medical avatars
US9931084B2 (en) * 2013-03-15 2018-04-03 General Electric Company Methods and systems for improving patient engagement via medical avatars
US20180168519A1 (en) * 2013-03-15 2018-06-21 General Electric Company Methods and system for improving patient engagement via medical avatars
US10913209B2 (en) * 2013-03-15 2021-02-09 General Electric Company Methods and system for improving patient engagement via medical avatars
US10546656B2 (en) * 2014-05-09 2020-01-28 Acupath Laboratories, Inc. Biopsy mapping tools
US20170352194A1 (en) * 2016-06-06 2017-12-07 Biodigital, Inc. Methodology & system for mapping a virtual human body
US10922894B2 (en) * 2016-06-06 2021-02-16 Biodigital, Inc. Methodology and system for mapping a virtual human body
US11291919B2 (en) * 2017-05-07 2022-04-05 Interlake Research, Llc Development of virtual character in a learning game
US11869150B1 (en) 2017-06-01 2024-01-09 Apple Inc. Avatar modeling and generation
US20200388363A1 (en) * 2018-02-26 2020-12-10 Children's Medical Center Corporation Extended reality medical report generating system
WO2019165432A1 (en) * 2018-02-26 2019-08-29 Children's Medical Center Corporation Extended reality medical report generating system
US11727724B1 (en) 2018-09-27 2023-08-15 Apple Inc. Emotion detection
US11801446B2 (en) * 2019-03-15 2023-10-31 Sony Interactive Entertainment Inc. Systems and methods for training an artificial intelligence model for competition matches
US11830182B1 (en) * 2019-08-20 2023-11-28 Apple Inc. Machine learning-based blood flow tracking

Also Published As

Publication number Publication date
CA2697309A1 (en) 2010-09-18

Similar Documents

Publication Publication Date Title
US20100257214A1 (en) Medical records system with dynamic avatar generator and avatar viewer
CN106909771B (en) For exporting the method and system of augmented reality information
US20190343717A1 (en) Device and method for three-dimensionally mapping acupuncture points
Gupta et al. Forensic facial reconstruction: the final frontier
CN109410337A (en) A kind of artificial intelligence medical system implementation method, system based on VR model
US11051769B2 (en) High definition, color images, animations, and videos for diagnostic and personal imaging applications
JP2018537782A (en) System and method for associating medical images with a patient
US20220375621A1 (en) Digital twin
CN113100935A (en) Preoperative puncture path planning method and training system for lung puncture operation
CN111096835A (en) Orthosis design method and system
CN110189324A (en) A kind of medical image processing method and processing unit
KR20200131020A (en) Medical machine learning system
CN109859823A (en) Calculate method, computing unit and the computer program product of individualized patient models
Marić et al. Facial reconstruction of mummified remains of Christian Saint-Nicolosa Bursa
Dao et al. Image-based skeletal muscle coordination: case study on a subject specific facial mimic simulation
Lindsay et al. Revealing the face of an ancient egyptian: synthesis of current and traditional approaches to evidence‐based facial approximation
CN112215969A (en) User data processing method and device based on virtual reality
Gualdi-Russo et al. Giovanni Battista Morgagni: facial reconstruction by virtual anthropology
Martins et al. InVesalius: three-dimensional medical reconstruction software
CN111523265B (en) System and method for reproducing cut injury cases
US20110242096A1 (en) Anatomy diagram generation method and apparatus, and medium storing program
Anastasi et al. Volume rendering based on magnetic resonance imaging: advances in understanding the three‐dimensional anatomy of the human knee
Sarvazyan et al. A new philosophy of medical imaging
Chen et al. A system design for virtual reality visualization of medical image
Kreutz et al. Forensic facial reconstruction-identification based on skeletal findings

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION