In GNUmed's field "Patient Request" I can enter the patients request individually and intuitive, like our diagnostic work is going on. This is what every GP is doing in daily work.
The other possibility could be: From here ub to the Field "Findings" the diagnostic strategy could be guided by a guidline-plugin, which leads to the Assessment. I do not mean a diagnosis automatism. This Plug-in-programm, that could do that job would be a big project of its own. Because the docters have various ways of thinking and because the diagnostic in General medicine is complicated (a never ending discussion-process) an adequate application will be a very big project. Nevertheless I am dreaming since years to get it realized. A guidline-programm for that purpose needs a data input mode, which does not disturbe the doctor/patient contact. The quickest way of data input in medicine is via keyboard, without mouse. But this means: both hands on the keyboard, concentrated on tipping into the keyboard. That is OK after having done my diagnostic chat with the patient. The patient does accept this interruption for administrative means. But hacking into the keyboard while speaking to the patient, with both hands on the keyboard, having a glimps over the shoulder to the patient is'nt what Balint would have done when speaking to the patient. So I think, for that part of dhe docter-patient-contact the better way of entering datas from the diagnostic chat into GNUmed would be to only (or predominantly) use the mouse. Do we need to enter data while doing the diagnostc/therapeutic chat? If we want our diagnostic chat beeing guided through a program, we have to enter data direct in front of the patient. This has to be done in a way, which does not or very few disturb the chat. How can we use the mouse to move thum-nail images of the symptoms of the patient's request on to an image of the human? It is not possibel in that simple way. But if we could divide and organize that process I have the idea, that we could manage it. This is, what I would like to test. The next question is: how can we provide guidance through the diagnostic process without a steep and inflexible sequence of question? How could it be that GNUmed knows now which filds I need to enter now? Which question have to be asked? When we are able to enter data while speaking to the patient, in a way, that does not disturbe the doctor-patient-contact, the Gnumed-programm could be able to think with us. Gnumed could be able to acompanay us in this diagnostic process (=Guidline). Gnumed should be able to offer fields to take up data whenever we need them. Like a good nurse during operation, knowing which instrument the docter will need now. A huge database would be necessary to save these diagnostic gatewayes. A self-learning programm would be necessary to learn these informations. Certenly this is a future project. But what I would like to start with now is to test, if the first approach is possible. As I am new in python and object oriented programming, I am not able to start this project on my own. Not now. But may be later :-) My question is: Does anyone know python source-code, that has realized to put graphical thum-nail-objects (could be the graphical sign for pain in our example) to some graphical surfice (could be the image of human beeing in our project). By positioning this object at a given position the following programmflow will depend on what is spicified on that position point. That means: The object "pain" will behave different on different locations. I would like to play with such a initial test-programm to see, if we can find a way of concentrating the many possible question at the beginning of a diagnostic chat to some few symboles. And I would like to learn object oriented programming with this example. Is this how the pink panther is thinking? Unrealistic? May be. But I am a phantast. And optimist. :-) Philipp Walderdorff _______________________________________________ Gnumed-devel mailing list [email protected] http://lists.gnu.org/mailman/listinfo/gnumed-devel
