Re: Image acquisition using video capture card
Thanx. The video output from the splicing machine is of NTSC standard and comes from a RCA jack. So can I use TV tuner card to save the picture on PC and Display it in LabVIEW? If I can't use IMAQ/Vision how can I see any picture in LabVIEW (linux)? Viji
Leaving cross marks and coordinate capture on mouse clicks in Vision
Hello: I need to customize an ROI to scan on a smaller area from a generated image. I wanted to leave cross marks on the JPEG image while I selectively select by my mouse and saving the coordinates on an XY-file. Any suggestions? Regards, Berns B. Bernardino J. Buenaobra University Research Associate National Institute of Physics University of the Philippines Diliman, Quezon City 1101 Philippines Email: [EMAIL PROTECTED] Tel/Voice: +632-434-4232 Fax/Data: +632-928-0296 Mobile:0916-2550-056 URL: www.nip.upd.edu.ph/ipl
Re: Problems using DAQmx Channel Property Node in Labview
Thanks Deborah It's good to know that the code does work. Mine gives the following error when I go straight into the channel property : code -200088 Property Node DAQmx Channel (arg 1) in add_units[1].vi append Encouraged by your success I'm going to go back into the MAX explorer and check through my channel set up to see if that's where my problem is. Thanks again for your help, I'll let you know how I get on. Ian
Re: Bird's Eye View 1.5.0 Available
Sheldon Instruments, Inc. [EMAIL PROTECTED] wrote: As they did the DLL utility 'downshift', NI will probably eventually incorporate this feature in their s/w It is part of LabVIEW 7.1, one of the few nice features you get with that versions. Although it can be sometimes discouraging to spend all the time to develop something like this, to just see NI to crank out the same, in a better or worse way, in the next version, it is definitely a very effective way to get new features into LabVIEW. Some sort of uber-suggestion reporting. Proofing to them it can be done, triggers some of their developers very effectively ;-) I had the same with my LLB viewer utility to show the contents of LabVIEW libraries in Windows Explorer ;-) Rolf Kalbermatter CIT Engineering Nederland BVtel: +31 (070) 415 9190 Treubstraat 7H fax: +31 (070) 415 9191 2288 EG Rijswijkhttp://www.citengineering.com Netherlands mailto:[EMAIL PROTECTED]
Re: Bug in edit items... of textpicture ring
Hello, this problem should be solved in the next LabView version. For the moment, there is no really good workaround. But I suggest to use a strict type defn control for this ring if it is used in several places in the program. This way you just need to update the original control and the changes will propagate throughout the VI. This is much easier than updating 3, 4, 5,... controls individually. Regards Thomas D. NI Germany (SRQ 203606)
Smart phones
Has anyone tried getting Labview onto smart phones running Windows Mobile? Seems to just be a special version of WinCE, and identifies itself as such if you dig down a bit. Being able to use a bluetooth-equipped phone wirelessly running the GUI of some embedded kit would be quite cool. -- Dr. Craig Graham, Software Engineer Advanced Analysis and Integration Limited, UK. http://www.aail.co.uk/
Re: Is anyone aware of a timing problem when using Notifiers ?
I think setting this parameter may help the issue but the initial issue regarding the removal of the sequence structure and the variable not clearing could be due to data flow. By using the sequence structure you are forcing the execution flow. I think what was happening is that you were getting a kind of race condition that either a delay or a sequence structure help solve the issue. Hope this helps, Steven Bird
Re: using snmp toolkit for labview
Hello Esther, I've searched and I'm afraid we don't have any example program on our databases. What I would do, -if you haven't done it yet-, is to contact Viodia to get help from them. Also, take a look at the LabVIEW ring as you may find additional forums and LabVIEW users. www.viodia.com http://c.webring.com/hub?ring=labview http://directory.google.com/Top/Computers/Programming/Languages/LabVIEW/ Regards, Jorge M. NI Spain
Re: using tab control on front panel to save space
Hello beam, you can use both versions. The advantage of the case structure is the avoidance of unnecessary processing of controls and indicators. This way processor resources are free for other things. In your example I would not use a case structure, as there are only two DBL values to check. If you work on large arrays a Case structure is helpful... With the case structure/tab control you can also check the user behaviour. Best regards Gerd
Re: How to multiply the measured signal with exp(-t/T) and integrate it?
Thanks much for your valuable answer. It gave me a good idea. However, I need more or less like a point-by-point type implementation to replace my measured pulse signal with its filtered value. Chunrok
Re: Fatal Internal Error:memory.cpp,line593
Hi Gerhard, I found an entry in our knowledgebase that should help you to localize the Error. Here it is: http://digital.ni.com/public.nsf/websearch/01B186B5CEB1D07786256747000F4319?OpenDocument Regards, Peter Weber NI Germany
creating and scrolling multiple histograms
Hi i'm working on a system that uses a VI to acquire data from brain research experiments, in which stimuli are shown to subjects during different trials. For every trial, a histogram should be plot conatining the nr of braincell spikes per time; So 16trials should give 16 histograms, and 100 trials should give 100 hisograms, but I do not know on beforehand how many trials there will be. And obviousely, I cannot display 100 histograms at once, so I was thinking of something like a scrollable pane that automatically extends when histograms are added, but I have currently no clue on how this should be done with a minimum of cpu load.. Any suggestions are welcome! Regards, Stijn
Re: Problems using DAQmx Channel Property Node in Labview
Hi there Deborah After finding no faults in my set up I re-installed my version of Labview and MAX Explorer. My program now works fine. Thanks for your help, and all the others who contribute to the group, my jobs been made much easier because of it. Ian
Calling Teststand from LabView
I've written a front panel GUI in LV 7.0. The operator presses a particular button to execute a certain test. How do you call a Teststand 3.0 seq ( that I've written) unique to that button when the operator presses it ? The LV Operator Interface shipped with LV was of no help.
Re: how to control rs232 signals (DTR, DSR, etc.)
Hi Vanessa, I recommend you to read this a href=http://digital.ni.com/public.nsf/3efedde4322fef19862567740067f3cc/1ecefaa7ad52994986256966005356ea?OpenDocument;Knowledgebase/a and to have a look into this a href=http://sine.ni.com/apps/we/niepd_web_display.display_epd4?p_guid=C3358273CB395047E034080020E74861p_node=DZ52363p_source=External;Example/a. Regards, Luca
Speech recognition using Labview 7.0 on winXP??
Hi all,I am new to Labview. I want to recognize my speech(voice),please give me answer of following question: 1.basic criterion by which I can distinguished different speeches(voices) and what is common in all voices for same character(ex. in a,e,i,o,u)??? 2.How to perform above task??(I mean rough Ideas like filtering,windowing,which plot frequency or magnitude or anything else,And Matching this means how can I say this is my voice not any one else??? Please help me I am really waiting for your reply so please reply me if you have any Ideas??
serial communication with Labview
I have to communicate with an old home-made device, via serial interface. In order to correctly open the serial port, in Qbasic, I had to use the following command: OPEN com1: 9600,n,8,1,CD0,CS0,DS0 FOR RANDOM AS #1. If I don't insert CD0,CS0,DS0 in my command line, I'm not able to communicate with my device. Does anyone know how I can do the same thing with Labview 6.1? I'v already tried the standard Labview functions, but they do not work. Thank you. Bye Carmen
Re: PXI or SCXI impedance analyzer?
Hi all, Previous message in this thread (in part): Subject: Re: PXI or SCXI impedance analyzer? From: Uwe Frenz [EMAIL PROTECTED] Date: Wed, 02 Jun 2004 10:05:15 +0200 George, you asked on 1 Jun 2004: I need to measure the impedance (low frequency, low power) of snip I do not really know what an impedance analyzer does other than measuring the resistance or impedance = 1/resistance of an object. IMHO a resistance measurement should be sufficient. snip HTH and Greetings from Germany! -- Uwe Frenz Uwe, et. al., Having invested 10 years of my career teaching electronics, I just had to throw in my two cents worth here. First however, Uwe, I must say that your posts are always excellent and I learn much from each one. Thank you for your contributions to this list. The comment impedance = 1/resistance is not correct. In fact conductance = 1/resistance, units Siemens (formerly mhos). Resistance is a DC parameter, i.e. the opposition to the flow of a DC current (mechanical analog might be friction). Impedance is an AC parameter, i.e. the opposition to the flow of an AC current (above mechanical analog extended might be stiction). Since there are an infinite number of AC frequencies it is more complex to measure. Theoretically impedance is generally thought of as being comprised of a DC (resistance) portion combined with an AC (inductive or capacitive) portion. To measure impedance one would need to measure the AC voltage across and the AC current through the device under test (DUT) and form their ratio. Note that these measurements would need to be conducted at every frequency of interest and that the voltage waveform and the current waveform would (for any not purely resistive DUT) have a phase shift between them. This phase shift is caused by the inductance or capacitance (termed the reactive component) in the DUT and must be measured as well as the amplitude of the v i waveforms. Since by Fourier analysis it can be shown that any periodic waveform is comprised of a weighted sum of sinusoidal waveforms at the fundamental frequency and integer multiples thereof, the above DUT would typically be excited by a purely sinusoidal voltage source. In time domain notation, v(t) = A * sin(w*t + theta), where A is the peak amplitude, w is really omega, the radian frequency of the sine wave, and theta is the phase shift. Dealing with this form of equation is challenging from the bookkeeping standpoint (read that pain in the rear) so this is where complex numbers are introduced to simplify the bookkeeping aspect. I won't go there as the proof is trivial and left to the student as an exercise. As a closing note, the reciprocal of impedance is admittance comprised of a conductance component (1/resistance) and a susceptance component (1/reactance, either inductive or capacitive). Having waxed sufficiently long-winded, I'll stop now. HTH and greetings from the U.S.A. Mark Watson Staff Engineer Philip Morris [EMAIL PROTECTED] (804) 752-5631 (804) 752-5600 (fax) (804) 215-5631 (pager)
RE: Programming approach
The mechanical engineer (ME) however decided that simply walking was sufficient to obtain the snack and much more easily executed than the first 10 foot jump. The same principal led this engineer to take up LV wiring and forego coding altogether. M -Original Message- From: William Rance [mailto:[EMAIL PROTECTED] Sent: Wednesday, June 02, 2004 11:37 PM To: Info LabVIEW Mailing List Subject: Programming approach Here's one which received some interesting solutions in python and c, wonder if someone would like to do it in Labview. A Pure Mathematician(PM) and an Applied Mathematician(AM) were put in separate rooms and located 20ft from an Apple.They were informed that to reach the apple they had to jump 10ft, followed by another jump of half the distance (5ft) followed by subsequent jumps each of which would be half the distance of preceding jump. The PM quickly calculated that no matter how may jumps he made he could never land at the apple, consequently he did not attempt the jumps. The AM however immediately started jumping, as he calculated that after 5 to 6 jumps he would land close enough to grab the apple. Happy coding to all. Bill
Re: Autoscale Y axis
I would like to autoscale the y-axis in a X-Y plot with the minimum set always to 0 and while the maximum autoscales. I tried to set y-scale start and minimum (using the XY plot properties) but still the y range minimum is set to the minimum value of the data. I would like the y-axis scale to be set at 0 for the minimum and the max. of the data for the maximum. Can anyone help with this. Thanks. You can either turn off autoscaling and set the max Y Scale each time you write data to the graph. There are some subVIs in the Picture Examples for determining the max and min using pretty much the same algorithm that LV uses. Or you can write the data to LV and immediately set the min to zero leaving autoscaling on. You might try this in a single property node or in a tightly sequenced set of property nodes, possibly with panel updates turned off when doing it. Greg McKaskle
Re: which command I can use to exit from a subvi without exiting
Similiar to 'Exit Sub' command in Visual Basic. There isn't a command to do this. Since LV is a dataflow language, it is important that all the data be delivered to the output terminals. BASIC is a procedural language, with all sorts of control flow commands like this, and if you are using a functional language or a dataflow language, you simply focus on getting to the end of the function. If you are in a while loop, you terminate the while loop by the value going to the conditional terminal. If you are in a For loop and need to exit early, perhaps it should have used a while loop. If there is code in a normal diagram that you sometimes do not want to execute, it should probably go in a case structure. The other reply has some good suggestions in that you probably have a UI or a sequence that you want to exit out of. Organizing your code so that you are in a button handling state or exiting state will help you organize how you think about the transitions between states. On the other hand, there lots of examples and user contributed VIs to look at. I'd recommend you get comfortable with a few of them since the state machine is just one programming tool, just like the loop and case statement that it is built from. Greg McKaskle
Re: Image acquisition using video capture card
Thanx. The video output from the splicing machine is of NTSC standard and comes from a RCA jack. So can I use TV tuner card to save the picture on PC and Display it in LabVIEW? If I can't use IMAQ/Vision how can I see any picture in LabVIEW (linux)? If you have the full version, you can use the picture control. If you have base, you can sacrifice some color information and use the intensity graph. Greg McKaskle
RE: Programming approach
I'll bite: An artist, an engineer and a mathematician were given the task of building a fence to contain a flock of sheep. The artist took into account the lay of the land and chose an appropriately aesthetically pleasing shape of fence, then painted it beautifully. The engineer counted and measured the sheep, then built a circle of fence to just fit around the flock. The mathematician put one fencepost in the ground and defined the rest of the meadow to be inside. PS. I've never met a mathematician (PM or AM) who could jump 10 feet! Ben Zimmer Software Developer, Meikle Automation Inc. PhD student, Applied Math, University of Waterloo. -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of William Rance Sent: Wednesday, June 02, 2004 11:37 PM To: Info LabVIEW Mailing List Subject: Programming approach Here's one which received some interesting solutions in python and c, wonder if someone would like to do it in Labview. A Pure Mathematician(PM) and an Applied Mathematician(AM) were put in separate rooms and located 20ft from an Apple.They were informed that to reach the apple they had to jump 10ft, followed by another jump of half the distance (5ft) followed by subsequent jumps each of which would be half the distance of preceding jump. The PM quickly calculated that no matter how may jumps he made he could never land at the apple, consequently he did not attempt the jumps. The AM however immediately started jumping, as he calculated that after 5 to 6 jumps he would land close enough to grab the apple. Happy coding to all. Bill
Signal Translation (Severely OT, with apologies)
Title: Signal Translation (Severely OT, with apologies) Hi I have a peculiar request - I want to take a 4-20 ma level transmitter signal and adjust it slightly by a non-linear equation based on another 4-20 ma signal. I can envisage several ways to do this involving custom hardware, some including a processor for the function, but I was wondering if anyone knows of a commercial product which would be capable of this off-the-shelf? Regards Blair Smith AECL CRL mailto: [EMAIL PROTECTED]
Re: daq invalid handle
Pourriez-vous pr=E9ciser votre configuration? Logiciel utilis=E9? Driver? Mat=E9riel? OS? Merci Fr=E9d=E9ric BOULLOT Ing=E9nieur d'Applications NIF
Re: how to convert time (X-axis) from seconds to minutes
I have Graph indicator and I am trying to change the units of time axis (x) from seconds to minutes. I wrote the small VI that increments the voltage at the output of the power supply in steps of 100 V every 1 min. After each time interval of 1 minute passes by the voltage is incremented by 100V. The resulting plot looks like stairs going up. Since my time interval will always be fixed to 1 minute I wish I can change the x axis so that it is scaled in minutes instead of seconds. So let's say I wanna go from 1 to 20 minutes on x axis in increments of 1 minute. Is that possible? What are the options I have? The graph is receiving numeric x data and when displaying in time format, assumes the data is in seconds. You have several options. If you are supplying a timestamp for each reading, multiply by 60 so that the plotted data is in seconds instead of minutes, then since that is how the graph interprets the timestamp, your scale can still be set from 0 to 20 minutes, and the graph will put the data where it belongs. If using a waveform, set the dT to 60. The second option is to continue using minutes in your timestamp, then on the graph either interactively or programmatically set the multiplier to 60. This was once in Scale Formatting..., and I believe in 7 it is in properties for the scale. Programmatically you have the same options of multiplier and offset. Set multiplier to 60 and the graph will label each unit you supply as if it were 60 units. For time, that makes seconds look like minutes. Greg McKaskle
Re: Autoscale Y axis
Hi Jeremy, Welcome to the Exchange! It is generally a good idea to summarize your examples that you post. That way; 1) We do not have to down load your code to see what method 1 is. 2) You can resize your porperty nodes and mainipulate the properties from top to bottom in the order you want. This will avoid the multi-frame seq structure. 3) Defering front panel updates and then undefering afterwards may improve performance and appearence. Trying to help, Ben
Re: Info-LabVIEW Digest - 06/03/04
Good Morning All, Hopefully this email will get through and not end up garbled as the digest version of my last post from this "Braindead AOL" mailer did. Unfortunately the only two email systems I have available on the road are AOL and Yahoo, and I _won't_ use the Outlook Express virus magnet that came on this laptop. I am having a problem with the digital I/O of a DAQPad-6020E initializing as all inputs, which apparently are floating at some voltage level around 1.3V. A 100K pull down resistor doesn't help and during the initialization time, while the DIO port init vi is in the lvdaq dll (about 3 - 4 secs) these pins are inputs, with the 1.3 v level, which causes at least one of the attached devices to turn on. I haven't had a chance to try the 10K resistors yet that were suggested by our Australian friend (I told my customer late yesterday that I would post last evening and probably have responses by morning, as "they are just getting up for work in Australia" !), but wanted to see if this post got garbled. The previous one was using AOL's web email access, this one is using AOL's own program. In neither is there an apparent way to turn off html or any other special coding. Unfortunately not all of us have the benefit of regular T1 or better access, and in some of the more remote locations that I work in I'm sometimes lucky to get good phone lines. Putnam Monroe
Re: Programme written in 6.1 is not writing to file in 7.0
Also, is there any other way I can use Labview 7.0 to write data acquired through a Hp8720 network analyser to a data file on the comp. Regards, Harshat
Re: Basics - edge detect or software counter
This just counts continuously as long as the button is held down. I could just hook the button to the loop enable and use the index as the count to do this. I need to count button pushes, need to detect the edge of the button signal. I tried using an AND to the loop enable, ANDing the button and NOT return from a boolean shift register (the shift register output is the button). This should allow it to run one loop cycle and then stop, but it doesn't work. Can you explain why? And in general, where is the detail on the loop and shift register operation? All the help and manual just give the basics, none explain why the above doesn't work. Is there a Labview for Dummies or something that goes into the fine details? Thanks-
Write Hexadecimal values as ASCII characters
Folks, How does one write a hexadecimal string as ASCII characters on to a file? I can display the characters on my front panel. But I don't know how to write it on a file. Any help is appreciated. V. P. Nair American Electric Power
Date format while writing to database using database toolbox
Hello All Again a problem with input locales and database tooldbox. How to write to database using connectivity toolbox in user defined format (not dependent on Windows locales)? Every time different user is logged on and writes a measurement to database, the dataformat is different depending on the user setting. I would like to define the date format in LabVIEW and use it always. Thanks in advance Pawel
Re: serial communication with Labview
I'm guessing CD is DCD, CS is CTS, and DS is DSR? In that case, these are inputs - what is Qbasic doing when you issue these commands/options?
Re: how to convert time (X-axis) from seconds to minutes
Thanks for detailed instructions. I am gonna try to implement it right away. Pedja
Re: Is neccesary to initialize an array of indicators to be able to show my results?
As your example shows with the indicators Array and Array2, you don't have to initialize an array to write to it. If, however, you use the Replace Array Subset, there has to be something in the array to replace before you actually replace it.
Re: non reentrancy for TCP/IP VIs.
The VIs and other discussion forum thread are here: http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=506500080024D8UCATEGORY_0=_49_%24_6_UCATEGORY_S=0 I would suggest continuing this thread there so as to avoid duplicates. Regards, Ryan K.
Error with Datasocket.vi
Hello, Due to a Mailproblem I repeat the inquiry. Sometimes I receive the error when i try to open a DataSocket connection (OPC). Error -2147417851 at DataSocketOpen and Error -1073479672 at DataSocketOpen No description found to this error. Any help or insight would be appreciated. Best Regards, Roland Känzig WIFAG Maschinenfabrik Wylerringstr.39, CH-3001 Bern Tel. ++41-(0)31-3393-476 Fax ++41-(0)31-3393-141 email: [EMAIL PROTECTED] Der Inhalt dieser E-Mail ist vertraulich und ausschliesslich fuer den bezeichneten Adressaten bestimmt. Wenn Sie nicht der vorgesehene Adressat dieser E-Mail oder dessen Vertreter sein sollten, so beachten Sie bitte, dass jede Form der Kenntnisnahme, Veroeffentlichung, Vervielfaeltigung oder Weitergabe des Inhalts dieser E-Mail unzulaessig ist.
I want to transfer data through the serial port in the same...
I want to transfer data through the serial port in the same coding that hyperterminal uses. How can i do it? The serial port seems to be working, and labview seems to be sending the data, but the problem is in which format does it send the data, because in hyperterminal i just input the string JDX and it sends it to my device, with labview it sends something but my device does not recognize it.
Re: How can I show/hide the windows cursor in a LabView 7.0 application run under Windows XP?
How are you doing it in Windows 2000? Are you making a system call to shut off the cursors? There is a palette of cursor manipulation (Application Control - Cursor or just do a palette search for Cursor). These VIs let you set the cursors, but I don't see an easy way to turn it off. A work-around might be to have a blank cursor in a cursor file and change the cursor to that. It will still have a click response (hot spot), so it does not keep the user from using the cursor, just seeing it.
Re: Is anyone aware of a timing problem when using Notifiers ?
Thanks for the response, the system works as expected now.
Thermocouple to Pt100
Hello I have a device which has a PT-100 temperature sensor. The probe of this is too thick for what I need. I want to measure the temperature of liquid in a test tube while I'm adding some reagent. It seems that the PT-100 is a device with some size to it, and is just too bulky to fit in the test tube along with a stir bar. I'd prefer to use a thermocouple. Does anyone know of a way to transform the signals from a thermocouple to mimic the a PT100. The Variomag reaction block I'm using has a very convenient way to handle this temperature (and I've already written the interface to it). Yours Sincerely John John Brohan National Instruments LabVIEW expert in Montreal Traders Micro We connect all sorts of things to computers 317 Barberry Place DDO Montreal PQ Canada H9G 1V3 Tel (514)995-3749 [EMAIL PROTECTED] http://www.TradersMicro.com/Index.htm
Re: Thermocouple to Pt100
John Brohan wrote: Hello I have a device which has a PT-100 temperature sensor. The probe of this is too thick for what I need. I want to measure the temperature of liquid in a test tube while I'm adding some reagent. It seems that the PT-100 is a device with some size to it, and is just too bulky to fit in the test tube along with a stir bar. I'd prefer to use a thermocouple. Does anyone know of a way to transform the signals from a thermocouple to mimic the a PT100. The Variomag reaction block I'm using has a very convenient way to handle this temperature (and I've already written the interface to it). Electrically you could do it by using the channel of a FET to simulate the PT100 and feeding the gate from the thermocouple voltage. With lots of extra bits around to tweak gains and offsets. It would probably be a pain in the arse to get right and make stable. But given it's possible there may be a module out there somewhere that does it, sold as an adapter or convertor, although I've never come across one. It would be a lot of work to simulate a 4 wire PT100 probe in software using a DAQ card. I've had custom thermometry probes made up by http://www.tc.co.uk/ in the past- they do a range including PT100 and may be able to help. Can't you use the temperature probe as your stirring stick? Or is a machine wiggling it? -- Dr. Craig Graham, Software Engineer Advanced Analysis and Integration Limited, UK. http://www.aail.co.uk/
check if printer is connected
hi, i am new in labVIEW utilisation, and i need to crate a vi who check if a printer is connected or no. i don't know how to start to create this vi and i didn't find example in LV or in teh code library on this forum. I am using LV6.1. can somebody help me? thanks in advance.
Keithley 2700/7708 module Measurement Function Error
I'm using Labview 6i trying to develop a simple VI to read a single temperature point from a type T thermocouple. When I set the Measurment Function to Temperature I get an error at the config block. The kicker is that this VI I have created is virtually a copy of an example VI included with the labView driver that came with the instrument. This example works fine. If I change the measurement function on my VI to DC volts, it works fine, just on Temperature that it fails. The error I get is Error -1074135024 occurred at KE2700 Configure Measurement.vi. Possible reasons: Driver Status: (Hex 0xBFFA0010) Invalid value for parameter or property. Secondary Error: (Hex 0xBFFC0002) Parameter 2 out of range, or error occurred while setting Parameter 2. Elaboration: Measurement Function I have attached my VI. Thanks Andrew
Re: serial communication with Labview
In your OPEN command, you are setting some of the handshaking things for a zero timeout, which you probably already know. Are you using the old Serial vi's (not VISA)? The serial port vi's have the handshaking lines disabled by default. I'd try this: Open up Serial Port Init.vi, inside there is Serial Port Reset.vi. From there you can Enable the handshaking lines RTS, DTR, CTS, DSR, etc. Perhaps just enabling those lines will fix you problem. It's probably do-able with VISA but I've never tried.
Re: serial communication with Labview
I'm sorry, Pawel, but I can't open your vi because my Labview version is 6.1. Please, can you save it as previous version? Thank you so much. Carmen
Re: serial communication with Labview
Maybe pawel can save his vi as 6.1. The property node is probably a better method. And VISA is always a better choice if you can use it.
Re: How to generate a continuous sine wave for a given amount of time
Thanks for the suggestion. Fortunately, I figured out how to solve my problem already.
Re: Basics - edge detect or software counter
br This is the problem when viewing the diagram versus actually using the VI. As I said in my first response, the Mechanical Action of the button is important -- it should be 'Latch When Released.' Right-click on the button, select Mechanical Action, and then select 'Latch When Released.' This will give you the desired behavior with my diagram. Try it. As for the in-depth fundamentals of G programming, in addition to the Online Help, and manuals, you may want to consider taking an Advanced LabVIEW course from NI. Here's the link for the training dept.: http://www.ni.com/training/ Can you attach your version of the VI so we can know why it isn't doing what you want it to do. Thanks, iKhalid/i br
Re: Write Hexadecimal values as ASCII characters
Hello V. P. [EMAIL PROTECTED] wrote: Folks, How does one write a hexadecimal string as ASCII characters on to a file? I can display the characters on my front panel. But I don't know how to write it on a file. Any help is appreciated. V. P. Nair American Electric Power There is a vi named Number to Hexadecimal String which you can find unde the string-tools strin/number conversion. I think this should do, what you want. Hope it helps Marco
Re: Thermocouple to Pt100
At 10:32 -0400 6/3/04, John Brohan wrote: Hello I have a device which has a PT-100 temperature sensor. The probe of this is too thick for what I need. I want to measure the temperature of liquid in a test tube while I'm adding some reagent. It seems that the PT-100 is a device with some size to it, and is just too bulky to fit in the test tube along with a stir bar. I'd prefer to use a thermocouple. Does anyone know of a way to transform the signals from a thermocouple to mimic the a PT100. Not really. A thermocouple generates its own voltage on a microvolt scale. A PT-100 is a resistor that changes its value which is about 100 at 0C. BUT you can get some small PT-100 sensors. Sensing devices has a pt-12A which we buy a lot of. These are about 1.6 mm in diameter and 12 mm long. I think they also have an 8 mm long one. http://www.thomasregister.com/olc/SmartCat.aspx?az=72980675type=details_orderbacklink=yestemplate=http://www.sensingdevices.com/template.htmcounters=noptno=PT100/12A -Scott
RE: Leaving cross marks and coordinate capture on mouse clicks in Vision
Berns, Monitor the IMAQ window for mouse clicks using Last Event. Whenever the mouse is clicked, add the XY coordinates to your list. Use overlays to add a small crosshair bitmap to your image. You can make a simple 5x5 crosshair and plot it at X-2, Y-2 using the bitmap overlay. Save the list to a file when you are done. Bruce -- Bruce Ammons Ammons Engineering www.ammonsengineering.com (810) 687-4288 Phone (810) 687-6202 Fax -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Berns J. Buenaobra Sent: Friday, June 04, 2004 7:25 PM To: [EMAIL PROTECTED] Subject: Leaving cross marks and coordinate capture on mouse clicks in Vision Hello: I need to customize an ROI to scan on a smaller area from a generated image. I wanted to leave cross marks on the JPEG image while I selectively select by my mouse and saving the coordinates on an XY-file. Any suggestions? Regards, Berns B. Bernardino J. Buenaobra University Research Associate National Institute of Physics University of the Philippines Diliman, Quezon City 1101 Philippines Email: [EMAIL PROTECTED] Tel/Voice: +632-434-4232 Fax/Data: +632-928-0296 Mobile:0916-2550-056 URL: www.nip.upd.edu.ph/ipl
Re: Write Hexadecimal values as ASCII characters
At 10:01 -0400 6/3/04, [EMAIL PROTECTED] wrote: Folks, How does one write a hexadecimal string as ASCII characters on to a file? format into string with a %X format (the create format string pop up will help you). other possiblities are %x, %0X, and %0x depending if you want leading radix indicators and your string in upper or lower case. Then write string to a file. Scott
Re: web publishing tool
Hello Ivan, You may want to take a look at this Knowledge Base entitled a href=http://digital.ni.com/public.nsf/websearch/E313D1D88CE14FDA86256E6A006B28ED?OpenDocument;Can A SubVI Open Its Front Panel When Using Remote Panels?/a As this KnowledgeBase explains, your subVI window should open up if you specify the subVI to open when called. You can do this by right-clicking on the subVI icon in your program and selecting SubVI Node Setup... and activating the Show Front Panel When Called checkbox. I hope this helps! Please let me know if there is anything further I can help with. Liz F National Instruments
Re: Upgrading LabView program from 3.0.1 to 6.1
Dan, Thanks for the info. This was very helpful. I'm going to find the conversion kit and give it a whirl. Thanks again! Dan
Re: Is neccesary to initialize an array of indicators to be able to show my results?
You could possibly use the Reshap Array function. In the example, if you first Reshape the Array and use 13 as the dimension size, then your Replace Array Suobset function would work correctly.
RE: Bird's Eye View 1.5.0 Available
Hi Michael and all Yes, LV7.1 has this feature and the scrolling is faster than PJM's utility. Actually, the NI version work in a somewhat different fashion. I think NI might have thought it better than I did. Instead of moving the whole diagram(or panel), they are just moving a visible area rectangle (representing the visible area of either the diagram or panel) over the diagram (or panel) image. Upon looking at it, I realized that it probably was a smarter idea as it simplify a lot of issue and will probably result in faster scrolling. I tried to see if I could port NI version to 7.0, and it did not work. Among other thing NI is calling library function nodes (within LV I believe) which might speed up the scrolling. Another nice feature of NI version is that it will update the picture whenever the diagram(panel) changes. In my last release (1.5.0) of the tool, I used so locked VIs from NI version to do the synchronized scrolling. Also, in the NI version, you cannot apply the viewer to various loaded VI's or VI's on disk. This is true, also Bird's Eye View is working for Labview 6.x and 7.x :) One thing I would love to see in the NI implementation is the ability to hook into the viewer programmatically. The whole NI package is located in: .LabVIEW 7.1 Root \resource\dialog\zoom_out.llb and I think the LV hook is the following VI: ZoomOut_LaunchZoomOutTool.vi PJM --- Michael Aivaliotis [EMAIL PROTECTED] wrote: Probably eventually might be quicker than you think. The link below lists the new features in 7.1, fourth on the list. And no, NI isn't just ripping off the idea, it has been in the works for awhile, but had to wait for some notifications before it worked well enough for us to support. Whichever you use, remember it isn't meant to encourage large diagrams. SubVIs are GOOD. So enjoy. Yes, LV7.1 has this feature and the scrolling is faster than PJM's utility. I don't mean to downplay the usefulness of the Bird's eye view utility. I think it is a great example of well written code. Also, in the NI version, you cannot apply the viewer to various loaded VI's or VI's on disk. One thing I would love to see in the NI implementation is the ability to hook into the viewer programmatically. This would allow the ability to program our own usefull tools around the viewer window. Michael Aivaliotis http://mmwis.com http://lavausergroup.org __ Do you Yahoo!? Friends. Fun. Try the all-new Yahoo! Messenger. http://messenger.yahoo.com/
Detective work on built executables
Dear built-executable gurus, I have an ugly labview problem. I have inherited a suite of labview programs, consisting of built executables that are known to work, and a vast, disorganised pile of vis. The challenge is to reconstruct working versions of the source code corresponding to each executable. What makes it hard is that there are 1/2 doz different versions of each vi, and no obvious way to tell which are the 'good' ones. I am hoping that it may be possible to dig into the exes and find out what vis went in to them (I understand that I certainly can't get the original source code out of the exes, but any clues would be helpful). Does anyone have any ideas? Thanks, Will Will Keogh Borehole Research Group Lamont Doherty Earth Observatory of Columbia University 61 Route 9W, Palisades NY 10964, USA Ph: 845-365-8673 Fax: 845-365-3182 [EMAIL PROTECTED]
Re: Detective work on built executables
Hi, On Thursday 03 June 2004 17:43, Will Keogh wrote: Dear built-executable gurus, I have an ugly labview problem. I have inherited a suite of labview programs, consisting of built executables that are known to work, and a vast, disorganised pile of vis. The challenge is to reconstruct working versions of the source code corresponding to each executable. What makes it hard is that there are 1/2 doz different versions of each vi, and no obvious way to tell which are the 'good' ones. I am hoping that it may be possible to dig into the exes and find out what vis went in to them (I understand that I certainly can't get the original source code out of the exes, but any clues would be helpful). Does anyone have any ideas? [...] Well, you can certainly try to rename the exe to .llb, then you can browse it with LabVIEW functions (List directory...). You could then try to open a ref to these VIs, and try to figure out the VI rev via property nodes (I really don't know if this is in the built app, though). Hope this helps, Joerg -- Leading SW developer - S.E.A GmbH Mail: [EMAIL PROTECTED] WWW: http://www.sea-gmbh.com
Prpblem with User Data Entry Into a Control
I'm writing some code where the user must correctly enter numeric data for a number of parameters (eg wavelength, time, energy etc) before the program continues (eg via a 'click button when done' type command). I'm using a while loop to lock the done button out and hold progress of the program until all data entered is valid (eg using a boolean AND statement across all data input parameters) For example, the entered time must between 30-60 seconds and a boolean test on the time will remain low until the time is within this band. Unfortunately, I've noticed that non-labview users do not click away from the control box (or use the tick box) when entering data, and a problem arises when they have establised valid data, but then start to enter stupid data in a control field (eg time = -5 seconds). If they don't confirm the data input (eg by using the tick box) my boolean trap stays high and the done button is available to allow them out of the loop. Unfortunately, once on the other side of the loop, the control carries through time = -5 seconds, which of course screws up the analysis!!! I have solved the problem in my own way, but my solution is not elegant, and users complain that the controls are 'clicky'. {And no, I can't tell them to enter the data properly!!!} Has anyone come across a neat solution to this problem?
Problem with User Data Entry Into a Control
I'm writing some code where the user must correctly enter numeric data for a number of parameters (eg wavelength, time, energy etc) before the program continues (eg via a 'click button when done' type command). I'm using a while loop to lock the done button out and hold progress of the program until all data entered is valid (eg using a boolean AND statement across all data input parameters) For example, the entered time must between 30-60 seconds and a boolean test on the time will remain low until the time is within this band. Unfortunately, I've noticed that non-labview users do not click away from the control box (or use the tick box) when entering data, and a problem arises when they have establised valid data, but then start to enter stupid data in a control field (eg time = -5 seconds). If they don't confirm the data input (eg by using the tick box) my boolean trap stays high and the done button is available to allow them out of the loop. Unfortunately, once on the other side of the loop, the control carries through time = -5 seconds, which of course screws up the analysis!!! I have solved the problem in my own way, but my solution is not elegant, and users complain that the controls are 'clicky'. {And no, I can't tell them to enter the data properly!!!} Has anyone come across a neat solution to this problem?
Re: pc-dio-96
Stu is correct is his explanation of the behavior. However, if the port is configured for output, you want to make sure you don't drive any of those lines with an external device. Otherwise, you can damage the board. It almost sounds like you are wanting to do a digital input operation with a set power up state. If this is the case, you can define what state your board powers up in. Refer to page 3-9 of the PC-DIO-96 user manual linked below. a href=http://www.ni.com/pdf/manuals/320289c.pdf;PC-DIO-96 User Manual/a
Re: Prpblem with User Data Entry Into a Control
SORRY I GOOFED THE TEXT ENTRY AND DUPLICATED MY QUESTION!! PLEASE DO NOT ANSWER THIS QUESTION, USE THE (CORRECT) ONE ABOVE. Perhaps a kind moderator will delete this duplicate - sorry!!
Re: Getting a parity error coming from VISA Read
Hyperterminal will certainly be helpful to make sure that you are getting valid data. You might also want to try VISA Interactive Control from Measurement and Automation Explorer under Tools...NI-VISA...Interactive Control. You will be able to exactly simulate the calls that are being made in your program such as writing and reading. Additionally, you will be able to see whether or not an error is returned from your read and what the error means. The basic problem here has got to be your stop bits, data bits, baud rate, or parity settings. VISA will not report a parity error unless the lower Windows API drivers do. Scott B. Applications Engineer National Instruments
Re: How to set the SE_SHUTDOWN_NAME privilege?
Thanks! That looks like it will do the trick!
Re: better method than using a large number of shift registers?
Hello Paul, Just a couple of thoughts as you continue to investigate your program. Your DAQ data may be gathered as a waveform data type, which combines data with timestamp information. If that is the case, you must extract just the data and ignore the timestamp if you are to plot data from different time periods on the same x-axis. So, you might wish to extract the data using the Get Waveform Components function found in the Functions Waveform palette. This will allow you to return just the array of data rather than the timestamped waveform. Then, you can analyze that array to see how many data points are being generated at once, etc. After you know that, you can implement a programming structure similar to mine because you will know how many times to call your DAQ code in order to generate the number of data points you want at a time (such as 100). If you'd like, post a screenshot of your code (or the code itself) and I'd be happy to help you sort through the DAQ programming, which sounds like the main stumbling block at this point. Hope this helps! Liz F National Instruments
Recursive Code Challenge (was RE: LabVIEW (list) recursion, or the SNIP elipsis are our friends ...)
Hi George, Yo' typed: ... but it is not allowed... try to drop a VI onto its own diagram and read the message! I was just making tongue in cheek humor :)) ... :)) Just teasing you mike! And I took it as such and am glad to see you did too. :)) I figured you probably knew about the VIServer method. Code Challenge: But I think that it will soon be possible to do exactly what you say above, grab a VI's own icon, drag and drop, and (if you have a switch or two set and some LabVIEW scripting in place) have the VI replaced on the diagram with a subVI created from a template (like the Factorial VI) modified by LabVIEW scripting, with the name of the VI replaced, or Run Method calls, etc. We just need to be able to capture the LabVIEW error/exception before the error dialog pops up and script away. Now that I think of it, I don't know of anything concrete, but I'd be pretty surprised if someone at NI doesn't either already have this or could code it up pretty quickly. Sounds like it would make a good IFLV/OpenG/LAVA/NIDevZone coding challenge considering that the scripting interface has been unofficially exposed recently. Any thoughts anyone? (or better yet, code) Ciao, Mike *** Michael C. Ashe Imaginatics Control Test Systems 11 Quinley Way Waterford, CT 06385 Phone: 860-444-2141 Cell: 860-961-0876 email: [EMAIL PROTECTED] ***
Re: Thermocouple to Pt100
Hi John MINCO makes some miniature RTD elements using thick films etc. I've recently used some of their polyimide encapulated ones. They apparently have a 316SS probe assy .125 in dia! See: http://www.minco.com/s602nr.php --BG Bill Gilbert, EM Tech UMN School of Physics and Astronomy Tel 612 624 4870 Fax 612 624 4578 John Brohan wrote: Hello I have a device which has a PT-100 temperature sensor. The probe of this is too thick for what I need. I want to measure the temperature of liquid in a test tube while I'm adding some reagent. It seems that the PT-100 is a device with some size to it, and is just too bulky to fit in the test tube along with a stir bar. I'd prefer to use a thermocouple. Does anyone know of a way to transform the signals from a thermocouple to mimic the a PT100. The Variomag reaction block I'm using has a very convenient way to handle this temperature (and I've already written the interface to it). Yours Sincerely John John Brohan National Instruments LabVIEW expert in Montreal Traders Micro We connect all sorts of things to computers 317 Barberry Place DDO Montreal PQ Canada H9G 1V3 Tel (514)995-3749 [EMAIL PROTECTED] http://www.TradersMicro.com/Index.htm
Recursive Code Challenge: (Part II)
Code Challenge: (Part II) grab a VI's own icon, drag and drop, and (if you have a switch or two set and some LabVIEW scripting in place) have the VI replaced on the diagram with a subVI created from a template (like the Factorial VI) modified by LabVIEW scripting, with the name of the VI replaced, or Run Method calls, etc. Just thought of a possible easier alternate method, the: General Recursion Call (Variant).vi 1. Use Call Chain or OpenG vi to get reference to calling VI. 2. Use LabVIEW scripting to get calling VIs connector pane. 3. Use Get Controls (Variant) from OpenG to read the indicator(s) of the calling VI. 4. Use Set Controls (Variant) from OpenG to write to the control(s) of the new instance of the calling VI. 5. Use Run Method with Wait for Completion = TRUE 6. Use Get Controls (Variant) to read the indicator(s) of the new instance VI. 7. Process and return... Granted, it is not grabbing own VI icon, but include in \user.lbi\... etc... Did I forget anything? Now, who will come up with the most elegant implementation, or a proof that this will not work? As my college prof's all said, the rest is left as an excercise for the student... Ciao, Michael *** Michael C. Ashe Imaginatics Control Test Systems 11 Quinley Way Waterford, CT 06385 Phone: 860-444-2141 Cell: 860-961-0876 email: [EMAIL PROTECTED] ***
Re: Saving the Control Values on a Front Panel
Matt, You can also use a datalog file. LabVIEW can write and read clusters to a file, and as long as the cluster don't change you should be fine. Basically your code would grab the value of each control, bundle it into a cluster and save that cluster to a file. For reading, you read your cluster from the file and get each individual value to drop it in the right control. I'm sure there are some shipping examples on this too. I hope this helps. Regards, Juan Carlos N.I.
Re: Prpblem with User Data Entry Into a Control
I'm not aware of any elegant solution to this, having come across the pronlem myself. A couple of suggestions though: rather than using logic to check the data range is correct, why not use the In Range coerce functions of the controls themselves, accessed by right-clicking on the control and selecting properties. This will ensure that only valid data is entered. Another alternative is to use a slide or knob control, again with the limits already set, or perhaps use a series of pop-up dialogue boxes prompting the user for data. Rather than having 'Next' buttons on the dialogue boxes, just ask the operator to press the enter key. You can then use a shift register to check that the contents of the control have changed before continuing.
Re: Obtain Queue.vi error 2 after long execution, Labview 6.1
Hi Jeremy, I agree with Rolf's suggestion of taking the obtain queue outside the loop. Better efficiency and better memory management were the reasons why I removed the Obtain Queue from the while loop in the VI I had attached. Please keep us posted on how the issue progresses. If none of this works, download the evaluation version of LabVIEW 7 (http://ni.com/labview) and see if that behaves similarly. Sincerely, Feroz
Re: Is neccesary to initialize an array of indicators to be able to show my results?
I do not understand why you convert the number to U8, leaving you only 8 elements for the resulting array if you later want to replace element #12. Whay don't you just remove the to U8 node, and you'll have 32 booleans in the resulting array to fiddle with. It does not make much sense to create an arbitrary length boolean array from an unsigned integer input with defined width. (If you only need 16 booleans in the array, replace it with To U16).
Re: New to active X, why does this not work
Which version of media player do you have? LabVIEW (6.1, 7.0, and 7.1) doesn't work with media player 9.0. a href=http://digital.ni.com/public.nsf/websearch/5EBD67575DD88DD186256E8B0050453E?OpenDocument;Windows Media Player 9 ActiveX Container Does Not Display Video in LabVIEW Front Panel/a Zvezdana S.
Re: How can I show/hide the windows cursor in a LabView 7.0 application run under Windows XP?
Hi, Take look at this example: a href = http://sine.ni.com/apps/we/niepd_web_display.display_epd4?p_guid=DA17D9A3808D6B73E034080020E74861p_node=%20DZ52102p_source=External;Confining Mouse Cursor to Front Panel of VI /a This does not hides the cursor, however you can limit it's movement to the bottom corner of the screen and the user won't be able to use the mouse at all. I hope this helps. Regards, Juan Carlos N.I.
Re: Autoscale Y axis
Hi Jeremy, I failed to mention that your option 1 is a good approach. If you look below you will see that Greg was on the same wave-length. Ben
Re: Upgrading from LabVIEW 6.1 to 7.1, get a MS C++ Runtime Error
I assume you get this error message everytime LabVIEW starts. This is not normal behavior. It is not even a common error. Have you tried reinstalling?
Re: VI hangs while attempting to save
Hi Deep, One thing you may want to check in particular is the number of controls on your front panel. Please take a look at the document below: a href=http://digital.ni.com/public.nsf/websearch/DECF7C29643B985886256C53004957F8?OpenDocument;Maximum Number of Front Panel Objects in LabVIEW 6.1/a I hope this helps. Sincerely, Feroz National Instruments
RE: Recursive Code Challenge
George Gatling wrote: ... I am truly amazed that a thread about commercial vs. free software and unsolicited advertising has morphed into a coding challenge to eliminate LabVIEW's recursion exception. Kinda reminds me of the whisper game we played in kindergarten! The whisper game was based on unintentional morphing, thats not the case here... As to solving the challenge... the obvious (to me) solution is to catch the exception and add the appropriate VI server code to the offending diagram. I would posit two choices, the VIServer code (more flexible, costs realestate) the General Recursion Call (Variant).vi compact... But I would bet a beer that as soon as someone does this, 7.2 will support recursion directly! :P Thats a fools bet, which I won't take. Did you know that PXI is a beer induced product? Legend has it that the initial conversation that kicked off the project came about on the roof of NI's HQ over a few beers at some sort of company micro-event. Never underestimate the creative potential of beer. At the Astronomical Telescopes conference in Hawaii two years ago, there was a paper/discussion on network and group/collaboration tools. Amoung this staid group of scientists, PhDs and general geniuses it was unanimously agreed by acclaimation that the number one engineering and scientific collaboration tool is ... yup, beer... On second thought, George, I'll take that bet Prost salud! Mike (sorry if this swerved off-topic, lets get back to code, beer later) *** Michael C. Ashe Imaginatics Control Test Systems 11 Quinley Way Waterford, CT 06385 Phone: 860-444-2141 Cell: 860-961-0876 email: [EMAIL PROTECTED] ***
RE: Suggestion for the list
To all Outlook users that this feature is not disabled by IT to change, Outlook HTML formatting can be turned off by two methods: After creating a new Individual message to the list: Select from the menu Format Plain Text Warning Dialog box Warning changing the format of this message from HTML to plain text will remove all formatting for this message . Are you sure you want to do this? Click Yes. For all outgoing messages: Select from the menu Tools Options Mail Format Tab Under message formatting Compose in this message format: Select the pull down: Plain Text Then select Apply or OK From: Bookwalter, Dan [EMAIL PROTECTED] Date: Wed, 2 Jun 2004 11:25:11 -0400 Lewis I am all too familiar with this feature, the list use to reject all my email and send me a nice notice with links to sites to explain how to fix the problem every time IT reconfigured Outlook, you wouldn't believe how hard it can be sometimes to get Outlook to send Text Only ALL the time. So maybe it can be turned back on. Jeff Christensen IP Core Gateways Motorola, Broadband Communications Sector [EMAIL PROTECTED] www.motorola.com/broadband
Re: VI hangs while attempting to save
FerozP, Just for completeness, could you tell us the related limitations in newer LabVIEW versions? Thanks!
RE: Detective work on built executables
Interesting problem. I just tried to use VI Server, but I get Error 15, resource not found for most VIs in the exe. It looks like only VIs that had their front panel built into the executable retained their version info. You can, however, get a list of all VIs in the executable which is better than nothing. Also, you don't need to change the name. I left the file as .exe and LV listed the VIs just fine. -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Behalf Of Joerg Hessdoerfer Sent: Thursday, June 03, 2004 11:58 AM To: Info LabVIEW (E-mail) Subject: Re: Detective work on built executables Hi, On Thursday 03 June 2004 17:43, Will Keogh wrote: Dear built-executable gurus, I have an ugly labview problem. I have inherited a suite of labview programs, consisting of built executables that are known to work, and a vast, disorganised pile of vis. The challenge is to reconstruct working versions of the source code corresponding to each executable. What makes it hard is that there are 1/2 doz different versions of each vi, and no obvious way to tell which are the 'good' ones. I am hoping that it may be possible to dig into the exes and find out what vis went in to them (I understand that I certainly can't get the original source code out of the exes, but any clues would be helpful). Does anyone have any ideas? [...] Well, you can certainly try to rename the exe to .llb, then you can browse it with LabVIEW functions (List directory...). You could then try to open a ref to these VIs, and try to figure out the VI rev via property nodes (I really don't know if this is in the built app, though). Hope this helps, Joerg -- Leading SW developer - S.E.A GmbH Mail: [EMAIL PROTECTED] WWW: http://www.sea-gmbh.com
Re: How to replace the logo of LabView at the top left corner of the program?
If you want to mess around with an executable that's already been compiled you can use a resource editor to change the icon or add icons with higher resolution and with more colors than the LabVIEW application builder will allow. One example of a resource editor is the PE Resource Explorer which is freely available for download (find it using Google).
[W] Machine Vision question
Hi, I have been asked to look into the feasibility of using Vision S/W and H/W inorder to automate verification testing of avionics boxes with Active Matrix or LCD displays. The display size is typically 4X5 inches. I think the requirement would be simple - take a snapshot of the display and get the text/numeric data from specific fields. This would be repeated for a number of different screens. I might have to verify that the data field is in the right location on the screen as well. Most of the box displays are monochrome active matrix Electro luminescent, but newer products are using color LCD displays. The resolution of the displays range from 320X192 up to 640 X480. I am a Labview programmer with 8 years of experience, but have no experience with machine vision. I would really appreciate it if anyone out there that has done something similar to this could provide answers to the following questions: 1. What type of camera would be required for this type of app, with what resolution? Do I really need a video camera or could I just use a single shot digital camera. 2. What NI board would I need. 3. How long would it take an experienced Labview programmer to come up to speed with the Vision Development module and OCR inorder to be able to produce a simple demo (grab a frame, read data from two or three fields)? 4. I am using Labview 6.0.2 - will I have to upgrade to 7.1 inorder to use the latest Vision Development Module? 5. What will the lighting issues be - since the display is back lit, would I even need additional light? Would I have to block outside light? 6. Would I also be able to perform optical testing of Luminance, chromaticity, legibility and pixel defects (not a requirement but for future growth ;) 7. What is recommended for fixturing. 8. Once up to speed with the vision tools, how quickly could a test for a new screen be done - hours, days?? I really look forward to the challenge of learning vision acquisition techniques, but I am a bit worried that I may have bitten off a little more than I can chew ;) Any help with this would be greatly appreciated, Howard Spec CMC Electronics Ottawa, Canada
Re: Error 6 occurs at GPIB Write
Error 6 is a GPIB Time Out. 1) Can the meter respond in 1 second? - if not, increase the timeout - P.S. I question why you put a clock wait in the same sequence as the read. Is it suppose to wait one second, and then read during the next, read then wait, or do those in parallel? 2) The INIT:CONT 0 in the 0 sequence doesn't have a line-feed/cr in it like sequence 1. Is that OK? 3) Does the instrument error queue (SYST:ERR? query) tell you anything - like bad command? Les Hammer
Re: Bird's Eye View 1.5.0 Available
However, with 'downshift' did the guys in Rochester receive anything for that? I hope so. Wasn't downshift the CIN to call DLLs? If so it would seem to me a long way between that and the current Call Library Node, especially considering that what that CIN did, was actually something which was really provided by the Watcom C runtime library used in LabVIEW all over the place. Rolf Kalbermatter CIT Engineering Nederland BVtel: +31 (070) 415 9190 Treubstraat 7H fax: +31 (070) 415 9191 2288 EG Rijswijkhttp://www.citengineering.com Netherlands mailto:[EMAIL PROTECTED]
RE: [W] Machine Vision question
Howard, 1. If you want to do color measurements, you need a color camera. You would need double the resolution of the screen (or more) to examine individual pixels. It sounds like you may need in the neighborhood of 1300x1000 resolution. You will need a little more resolution to handle the different sizes of screens and variations in position of the screen. I strongly suggest a camera that converts to digital onboard instead of sending an analog signal. You also need to select the appropriate lens to give you the best field of view. Graftek Imaging (800) 441-2118 is a good place to get the camera, lens, etc. They can help you figure out what you need. 2. I prefer firewire for its simplicity. The actual connection to your computer will depend on the camera. You just need to get the board that matches your camera type. 3. This could possibly be done using Vision Builder for Automated Inspection. It wouldn't take long to get a prototype working. I'm not sure if the final version would work in VBAI or if it would need to be converted over to LV. 4. No, you don't need to upgrade. You might be missing a few new tools in the old version of LV, but the newest version of Vision will work with older versions of LV. 5. Good question. If the backlight isn't strong enough, you would need additional lighting. If outside light is on the same order of magnitude as the backlight, you will need to shield the outside light. Lighting is a little bit of trial and error. 6. With a high resolution color camera, you could probably do most of this. If you go with a less expensive BW camera, you probably won't be able to do this. 7. Fixturing really depends on the system. You don't want the camera to move relative to the screen so that the focus and field of view are consistent. It also depends on what lighting is used. 8. It really depends what the test is. If you are just adding another of the same, such as OCR of certain regions, it would be hours. If you are developing a new test to detect missing pixels, it would take much longer. The most critical thing in vision applications is getting a good image for analysis. Camera, lens, lighting, field of view, depth of field, etc. all play a part in this. Once you have a good image the analysis can be fairly straightforward. Also make sure you have enough resolution for the OCR to work. The width of each line in the characters should be several pixels in your image for the best results. You will need to do a lot of prototyping to make sure this works properly. Bruce -- Bruce Ammons Ammons Engineering www.ammonsengineering.com (810) 687-4288 Phone (810) 687-6202 Fax -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Spec, Howard Sent: Thursday, June 03, 2004 2:21 PM To: Info LabVIEW Mailing List Subject: [W] Machine Vision question Hi, I have been asked to look into the feasibility of using Vision S/W and H/W inorder to automate verification testing of avionics boxes with Active Matrix or LCD displays. The display size is typically 4X5 inches. I think the requirement would be simple - take a snapshot of the display and get the text/numeric data from specific fields. This would be repeated for a number of different screens. I might have to verify that the data field is in the right location on the screen as well. Most of the box displays are monochrome active matrix Electro luminescent, but newer products are using color LCD displays. The resolution of the displays range from 320X192 up to 640 X480. I am a Labview programmer with 8 years of experience, but have no experience with machine vision. I would really appreciate it if anyone out there that has done something similar to this could provide answers to the following questions: 1. What type of camera would be required for this type of app, with what resolution? Do I really need a video camera or could I just use a single shot digital camera. 2. What NI board would I need. 3. How long would it take an experienced Labview programmer to come up to speed with the Vision Development module and OCR inorder to be able to produce a simple demo (grab a frame, read data from two or three fields)? 4. I am using Labview 6.0.2 - will I have to upgrade to 7.1 inorder to use the latest Vision Development Module? 5. What will the lighting issues be - since the display is back lit, would I even need additional light? Would I have to block outside light? 6. Would I also be able to perform optical testing of Luminance, chromaticity, legibility and pixel defects (not a requirement but for future growth ;) 7. What is recommended for fixturing. 8. Once up to speed with the vision tools, how quickly could a test for a new screen be done - hours, days?? I really look forward to the challenge of learning vision acquisition techniques, but I am a bit worried that I may have bitten off a little more than I can chew ;) Any help
[OT] NI Internal Programs
I've always wondered what NI uses internally to write LabVIEW. To be more specific, the LabVIEW core (not RT, PDA, etc.) on Windows. What is the language breakout? (%C, %C++, %Assembler, %LV, etc.) Which development environment? (MS Visual Studio, Borland, GNU??) Any third party tools? (compilers, editors, debuggers, etc.) Bug Tracking tools? (Bugzilla, etc.) SCC? (SourceSafe, CVS, etc.) Patrick Lavezza [EMAIL PROTECTED]
What is the best way to update a dynamic vi?
I am loading charts and graphs dynamically and I want to be able to update them using the same loop in the main vi. (Using LV 6.1)
Does anyone have drivers for the HP 16510 logic? analyzer
I have driver for the Mainfram 16500,but looking for module drivers for the 16510
Bessel function accuracy
From a side-by-side comparison with results generated by both Maple and MatLab, I realized that although the output of the VI evaluation of Bessel J function is a double-precision array, the accuracy is only up to single precision. Is there any simple way to increase the accuracy to double precision?
Is ther a patch to allow using Vision Builder 6.1 with LabView 7.1?
I am currently using Vision Builder 6.1 with LabView 7.0.(thanks to a patch I found here). Is there a simular parch for LabView 7.1?
Re: Problem with User Data Entry Into a Control
Thanks Jean-Pierre, Excellent. I think this will do the trick; I've tried just about everything else (including local variables and tunnels etc), but I think your solution is the one that will do it. I'll give it a go. Incidentally, I couldn't use the coerce option because I wanted to avoid forcing the user to any particular data value, especially extreme ones at either end of the range!! Thanks Blueside
Re: Prpblem with User Data Entry Into a Control
Thanks SLH, I'm going to go down the shift register route - the only one I didn't try!! Your comment on slider inputs is noted - I've found users really love to enter data this way, especially when the analysis is interactive. thanks Blueside
Windows Media Player runs when VI Is opened
I have a VI that now runs nicely to play an audio file. The only disturbing feature is that when I open this VI (but it is not running), the player launches. I though I could fix by putting in a subVI, in a case that required a true inputWrong, crazy thing still launches. In the actual VI, I close the player properly (player says closed) so not sure what else to try.
Re: image background in the front panel , is it possible ?
go to Paint (or your favorite drawing program) select and copy the image (this puts the image onto the clipboard), go to labview and paste the picture on the front panel. You can also attach an image to a control (like a button) by right clicking the fP object, selecting Advanced Customize, and (in the control editor that is launched) right click the object and select import picture. This method is nice because you can import one image for the true state, and another image for the false state. I've used this method for small push buttons to show that the button was pressed. For example, for a Load button, you might have the traditional windows load icon (a folder) and when you press the button, the folder opens.
Re: Error 6 occurs at GPIB Write
As it turns out the problem was the MEAS command isn't what I thought it was. So I was able to fix that problem. The clock wait itself I removed from the program as it was just a debugging method. I am still getting a weird reading error though were for some reason the gpib card will return a string of ++6.17345E-10 instead of +6.17345E-10 could this have something to do with me using a PCI GPIB card I bought a long time back from a 3rd party vendor? (MeasurementComputing.com)?
Re: What is the best way to update a dynamic vi?
Use a Call By Reference node to pass the data directly to the dynamic VI. There is a great shipping exaple that illustrates how to use the call by reference node. Launch the example finder go to the Search tab type in call by reference click the Call By Refernce Nodes text that apperas in the search results field double click the example called Dynamic Load example.
Re: creating and scrolling multiple histograms
I'm not sure if this will work for you or not, but this is what I thought of when I read your post... You can't look at 100 histogram simultaneously. You can probably only fit 4 or 5 on a front panel (without using tab a control, and I recomend NOT using the tab control), so I would try to work with a front panel that contains a (small) limited number of Histograms. You should already be saving your data to a file, so I would create a new directory each time a new set of trilas is started. Then create a data file for each trail you run. So in the case you described, the directory would have 16, or 100 files. Then on the front panel, you can have a ring control under each histogram that lists all the files in a given directory. When the user selects an item from the ring control, you load the data file, and push the data into the histogram. This would let the user quickly select the data run they care about, look at it, and select another.
Re: How would you recommend that I schedule a file-creation...
To avoid polling you could calculate (in a separate loop) the amount of time left before the next save and then feed this time to the timeout of a wait notification node. When the notification times out, write file. Other processes can control/abort this loop sending notifications to the loop e.g. when the wait notification node returns, you check for notification data or a timeout.