Re: Dual monitors for labview use
Scott, you asked on Thu, 12 Feb 2004 13:26:27 -0600: Can anyone tell me what graphic cards work well with respect to using dual monitors and labview? How does the graphic card handle moving the display data from one monitor to another. I was considering the ATI Radeon 9800XT. Great for gaming as well. I remember someone talking about having the diagram window on one machine and the panel window on another. No more fussing around with the windows to do your work faster. Well, here at getemed most of the devellopers have now a dual monitor system. Most of us have matrox G550s, but almost all dual monitor cards should work well too. I had a project where the customer even wanted three large monitors. So I added two graphics cards, an ATI and a Matrox dualhead into the machine. All worked well, allthough (because of the 3 monitors) I had to use Win98 at that time. Here we come to a big caution: Win NT 2k (no knowledge on XP yet) handle multiple monitors as two windows to a large and unique desktop, whereas Win9x used to have multiple independend desktops that just had to share one side with each other. This might be important, because almost all dualhead cards I know of have an excellent head as primary head, supporting video bandwidth of up to 250 MHz (and resolutions of up to 1600*1200). The secondary head is propably less capabal, supportimg only smaller resolutions. This may result in some area on the desktop that is not displayed on any of the screens in W2k. Example: One screen with 1600*1200 left and another with 1280*1024 right aligned at bottom. The desktop will be (1600+1280)*1200 and there will be an area of (1600, 1199) to (2879, 1024) that will be invisibel. On Win9x however there will be just two independend desktops. Also, how about those kvm monitor switches? Anyone used those in conjunction with the dual monitor graphic cards? Is going with an extra PCI board as well as the ATI Radeon 9800XT AGP board the better route than running both monitors off of the same card? Thoughts? Comments? Monitor switches have to support the high video bandwidth in order not to damage signal quality. Try it out for the equipment and resolution you are trying to use. One other off topic question. Have you guys/gals been buying and using LCD monitors? Anyone running them for a long time? I keep seeing used monitors show up on ebay that have a few/many lcds burnt out or broken in the display. How hard and expensive are they to repair if this is the case? I can solder as I'm a EE by nature. AFAIK this is an internal damage that can't be repaired and never ever with a 'burning iron'. Nothing to say about your 'natural' solder art crafting. (Hope this is got as that joke it is intended to be. Joking in a foreigh language is not so easy.) The LCD producers have even defined a level of defective pixels that are considered as OK according to the production standards. Otherwise they would get too few panels being OK out of their production lines and those would be too expensive for mass market. Greetings from Germany! -- Uwe Frenz ~ Dr. Uwe Frenz Entwicklung getemed Medizin- und Informationtechnik AG Oderstr. 59 D-14513 Teltow Tel. +49 3328 39 42 0 Fax +49 3328 39 42 99 [EMAIL PROTECTED] WWW.Getemed.de
Re: Vi converted to LabView 7.0 Error
Hi Dr. Imad, another workaround is described a href=http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800939BUCATEGORY_0=_49_%24_6_UCATEGORY_S=0;here/a. Search for my answer. This is a conversion bug in LV 7.0 and the workaround is first load it into LV 7.0 then save it into the new location and then make your changes. Waldemar
Re: Dual monitors for labview use
Hi Scott, Can anyone tell me what graphic cards work well with respect to using dual monitors and labview? How does the graphic card handle moving the display data from one monitor to another. I was considering the ATI Radeon 9800XT. Great for gaming as well. I remember someone talking about having the diagram window on one machine and the panel window on another. No more fussing around with the windows to do your work faster. We use dual monitors here, using a variety of different graphics cards. I have an nVidia GeForce FX5200 (great card, but take care if you buy one because they do single- and dual-monitor varieties), my boss has a Matrox card of some description, and one of our machines just uses two separate single-display cards. That machine runs XP, which has dual-monitor support built in, so you can use two separate single-display cards instead of a single dual-display card. In all cases, LabVIEW works just fine with them. The software we develop uses a subVI to retain the screen position of VIs and put them back in the same place next time they're launched (using the VI's Windows.Bounds property), and that has no problem coping with the dual display setup. As someone else has noted, XP and Windows 2000 treat dual monitors as one large desktop, so software doesn't need to be aware of which monitor it's running on, it just runs somewhere on that large desktop and neither knows nor cares whether that happens to be on monitor 1 or monitor 2. The only noticable difference between the various solutions we use is the facilities offered by the driver software that comes with the video card(s). My nVidia card comes with an app called nView that allows you to do all sorts of fancy things with your dual monitors, such as enabling window spanning across desktops, choosing which monitor specific apps open on, setting different wallpapers for each monitor and so on. The machine that uses two separate cards and relies on the dual monitor support built into XP lacks these advanced features, although it's still perfectly usable. If you have an XP machine with a spare PCI slot, a spare PCI graphics card and a spare monitor, try plugging the card into the PC and giving it a go - won't cost you anything, and you'll see it first-hand. Try before you buy! :-) One other off topic question. Have you guys/gals been buying and using LCD monitors? Anyone running them for a long time? I keep seeing used monitors show up on ebay that have a few/many lcds burnt out or broken in the display. How hard and expensive are they to repair if this is the case? I can solder as I'm a EE by nature. I've got an LCD monitor as my 2nd display - a fairly cheap 15 (1024x768) model that I have pivoted through 90 degrees (another feature supported by the nVidia software), making it great for working on documents. I've had it for about 4 months and haven't had any problems so far with dead pixels. We've got another LCD display that we've had for about a year, again no problems with dead pixels. Not sure how feasible it is to repair a dead pixel. Note that even new LCD monitors may ship with some dead pixels, and most manufacturers specify a tolerance level for dead pixels below which they won't replace a panel. All the best, Simon Whitaker [EMAIL PROTECTED] Software developer, Tiab Ltd tel: +44 (0)1295 714046 fax: +44 (0)1295 712334 web: http://www.tiab.co.uk/
motion control feedback (analog) PID
Hi there LV'ers I have a few motion control problems that are, to quote ' doin my head in'. I have a system that uses both encoder and analog feedback at varying times during the program operation. This all seems to work fine with the system configuring and re-configuring the analog/encoder etc what ever is required. My system is then to perform two task under analog control, one is to simply drive to an analog feedback value and hold, the other is to cycle between two analog feedback values. The main part of the problem that I have is the setting of PID parameters using the MAX autotune facility. This autotune feature works fine with encoder feedback with values in the range of 10 .. 250 for the 3 main PID parameters. When this same task is performed with analog feedback the machine drives at high velocity and comes in with values of 32767 for the PID parameters - clearly an error. If we try to set the PID parameters for analog feedback manually we come up with a few values in the 10 .. 100's range for the parameters and the motor sort of behaves itself except when we try a simple cycle between two values in the MAX 1-D interactive screen. The one direction move say to 200 (whatever's) works fine, a cyclic move from 200 to 600 will cycle the motor very hard and not actually stop on the desired values etc. These problems seem as though they are all PID problems and I just can't work out why I can't auto tune on the analog feedback. The transducer we are using is a load cell, although I have changed this to an LVDT just to ensure there wasn't a noise problem on the signal. Thanks for any help wayne -- Wayne Galbraith School of Civil and Resource Engineering The University of Western Australia 35 Stirling Highway Crawley 6009 Tel - (08) 9380 3788 email - [EMAIL PROTECTED] --
Re: network oscilloscope
Matthew, thanks for your answer. The Scope Chart Mode is very interesting.. I actually wanted to know how to implement the way a DSO works in LV (setting trigger level and timebase etc.), but this isn't necessary anymore since I decided to use my dsp to calculate the data to be plot, hence acting as a DSO. Then it sends this data to LV, so LV should only display it, and I'm probabely going to use the Scope Chart for that. -Stijn
I need that Programm for LV 5.0
thank you jonnie 5 but there are only programs for LV 6.0 can I download programs for version 5.0 or a LV-Update
Re: How to compile wavread on MatLab Script?
When you say it will not compile, do you mean you've got a broken arrow or you get an error? If broken arrow, what does the error list say the problem is? If you're getting an error out of the Matlab Script structure, what is it? I've tried this on LV 6.1 and Matlab 6.5 and it works fine for me. Post your code we'll try to help. Tim
Re: How to compile wavread on MatLab Script?
Also, check out this link - don't know if it addresses your problem. http://digital.ni.com/public.nsf/websearch/4475BC3CEB062C9586256D750058F14B?OpenDocument
Re: DAQ+storage must happen independent of other screens OR user interaction
There are some examples how to use named queue. LV 6.0.2: Search Examples Advanced Examples Execution control Examples Synchronisation examples Named Queue LV 6.1 LV 7.0: Find Examples Optimizing Applications Synchronizing Tasks Queue Message Loging.vi In this example they use multpile sender and one receiver, but they show using a named queue to communicate between parallel running VIs. In the same example categories are also examples for notifiers who demonstrate how to send data from one sender to multiple receivers. I hope this will help you but feel free to ask again. Waldemar
Re: Optimizing network throughput over Ethernet with TCP/IP
Ethernet packets are range from 46-1500bytes, or there abouts, so anything is being broken up. Well, anything over 1500 bytes. I'll give the even multiples of 32768 a try and see how that goes. (Well, there seems to be support for frame sizes up to 9000 bytes in Gigabit Ethernet.) I'm going to have to experiment a bit more to see whether the delays are caused. I may also try to create an app in C or find a test app that will pump data through, to test out LV's receive limitations. Adam
Re: Why does converting a SGL to SGL get rid of coersion dots?
Well changing the setting and re-opening LV is not enough to fix this issue. Since the wire has to be dis-connected and then re-connected to get rid of the coercion, this has to be an issue with what LV is doing at edit time. So... Ben
RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!
Whos the darn fool at NI that set the default mode for numeric displays to Hide Trailing Zeros. Try setting the displays to digits of precision instead of significant digits The other annoyance (for DBL controls and indicators) is that they set the default formatting to automatic with 6 significant digits. How many times do you want to see 6 digits after each number? I'd like to see it set to floating point with 2 digits of precision. I suggest to NI that a method be included that will allow the LabVIEW programmer to set these defaults in LabVIEW for all controls and indicators globally.
i created boolean references in my main vi block diagram and...
i created boolean references in my main vi block diagram and copied them to my sub vi front panel. when wire my reference in my main vi to one the input node of the sub vi the wire is broken. the error says its a class conflict why? i created boolean references in my main vi block diagram and copied them to my sub vi front panel. when wire my reference in my main vi to one the input node of the sub vi the wire is broken. the error says its a class conflict why?
Re: Clusters: How to remove the 3D shaded background?
Well here is a golden nugget hidden at the bottom of a pile of fluff! Transparent Front Panle Backgrounds! I wounder what I could do with that? Ben
Re: Dual monitors for labview use
Uwe Frenz 02/13/04 02:49AM ...but almost all dual monitor cards should work well too. ... This may be true in general, although I saw one dual monitor card do nasty things. The brand name was something like Apien or Apex or something like that. (does that sound familiar to anyone?) It was unbelievable how badly this card performed. Certain everyday tasks would only work on one monitor, and there were other problems I forget now. I am using a Matrox G550, and it works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED]) John Howard
Re: divider, advice needed, please
Hello guys; Firstly let me thank a few people who answered my question, regarding dividing a signal freq. The funny thing is that the whole concept looks so easy and so trivial, but IT DOES NOT WORK! All the people who answered me told me to use Down Counter or Divider Config.Vi I can swear to God that I have done this but it did not work and that is why I asked the list for some help. After all I don't think I am that daft! But I can tell now that I am stuck now. So, the question is Has any body used the Down Counter or Divider Config.Vi, successfully? Do you know where can I get a working example? I am going to ask NI because this looks to be a question for them Thanks again + Dr. Hamid R. Yazdi Federal Mogul Manufacturing technology 3935 Research park drive Ann Arbor, MI 48108 Tel: 734 222 4108
Re: Problem with DAQMX digital output
Are you using the daqmx functions? If so please post the code. Or you can try using the Daq assistant. This will step you through the process of configuring a digital output. hope this helps.
Re: How can i display the signal out of a FFT block to a graph?
Hi, For the Y-axis, it depends of the input data binaries. I mean, if these come from a voltage signal or a collection of data or a sound signal, etc. It's necessary to know the factor conversion the voltage values, for example... And for the X-axis, that represents the axis of frequencies, and its posible to reconstruct until the nyquist order of the harmonic frequency (FS/2).Also it's important, the lenght of the input data, it's a power of 2, since algorithm FFT gives a fast and optimal result. Then connect the output to a Waveform Graph Display, on the front panel. I hope this help. Bye..
Re: Clusters: How to remove the 3D shaded background?
Don't you just hate it when that happens? :) Eric
Re: Filtering waveform?
How can i do re-sampling in LabVIEW 6.1?
RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!
Jack, I have started playing with LabVIEW 7 recently and this is already starting to drive me batshit. My high school science teachers would dock marks for failing to properly account for significant figures. To a physicist (and I assume for other technically inclined professionals) 0, 0.00 and 0.00 are three *different* numbers. And I haven't even specified my uncertainties... :) Excel also has this annoying default behaviour. Sincerely, Johann Junginger. -Original Message- From: Jack Hamilton [mailto:[EMAIL PROTECTED] Sent: Wednesday, February 11, 2004 12:50 To: LabVIEW -Info Subject: LabVIEW 7 Numeric Display precision - or Please Kill me now! Who's the darn fool at NI that set the default mode for numeric displays to Hide Trailing Zero's. I am now getting carpal tunnel just correcting this every-single-time I drop a numeric display on the front panel. Why is God name would I declare a single or double precision number - and NOT want to see the values to the right of the decimal? Nearly without exception clients want to see the darn decimal places. And don't start that modify your Ini line - I work on many different computers and can't go customizing the modifying every single one. Does this seem like a rant - Yup ;- Fed by the 100's time I've had to change the display configuration - I can do it now with my eyes closed. This is like may SUV, the back door swings the wrong way - now I have to stand in traffic to put stuff in the back - gee seemed like a good idea in the development lab... Jack Hamilton Hamilton Design [EMAIL PROTECTED] www.Labuseful.com 714-839-6375 Office
Re: Dual monitors for labview use
John, you wrote: I am using a Matrox G550, and it works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED]) Thats about or little above the specs, AFAIK. The highres requires a pixel clock of 160 MHz. The video bandwidth of that chips is not far beyond 250 MHz, so the signal quality will not be very impressive. But it works - i use exactly the same resolution on a Nokia 445Xpro. But the second head used to have a much smaller video bandwidth. The producers tend to overstate their product capabilities, e.g. allowing resolutions that do not produce good signal quality - just to have impressive marketing labels. -- Mit freundlichen Grüßen! Greetings from Germany! -- Uwe Frenz ~ Dr. Uwe Frenz Entwicklung getemed Medizin- und Informationtechnik AG Oderstr. 59 D-14513 Teltow Tel. +49 3328 39 42 0 Fax +49 3328 39 42 99 [EMAIL PROTECTED] WWW.Getemed.de
conflicts between property nodes
I am using property nodes on image display to set the position center ( X and Y ) and the zoom (X and Y). When using only position center property nodes everything works fine , but if i set the zoom value with property nodes the position center doesn't work any more, the image stays in the default position. Is there any known reason for this behavior? I am using Laview 7 under windows 2000 professionnel, and i am acquiring an image with a PCI 1408 card. Thanks,
Frequency spectrum of IFFT output
Output of IFFT block is summation of N orthogonal sinusoids. These sinusoids each have different frequency and the lowest is 0 (DC). In reality output is array of complex numbers which represent complex weight for corresponding sinusoidal basis function. I neeed to know how to present output signal in frequency domain spectrum. I'm confused with complex numbers. Anybody knows solution to this problem?
Re: Problem with array processing !!!
It's due to the fact that you are displaying all those 100 000 values and updating them in the indicator on every iteration. If you move the indicator outside of the loop so that it will only be updated when the processing is finished things will run fast.
Re: Problem with DAQMX digital output
Yes, the 6035 is supported (unless you're running win 95... or 3.1... you are on windows, right?) See: http://www.ni.com/support/daq/versions.htm The troubleshooter reffed above is at: http://www.ni.com/support/install/
Re: unable to read the frequency from this output sine wave of amplitude VS time
I don't really see the utility of this VI, however, the variable scan rate, of the LJstream graph processing, it's an arbitrary value, why?... What is the idea with these VIs?, tell me more, please. Bye.
Re: mouse up event on a slide
Sorry, I misunderstood your problem. I was able to re-create your problem. This could be considered a bug, moving off slider before letting mouse up should return the pointer to its original value, but it does not. So now your event doesn't fire because it never saw a mouse up event on the slider. The only fix I could see is to set the event trigger to Value Change. Yes you will have an event for each incremental change. Don't know of a good solution.
Re: Serial comms
At 12:30 + 02/13/2004, Craig Graham wrote: The next is in the other port settings. The manual specifies 1 start bit, 8 data bits, 1 stop bit, no parity. The VISA serial stuff in Labview 7, which I've only just started playing with, has properties for stop bits, data bits and parity which I've set, but I can find no option for start bits. Is this a problem? On the offchance, I've tried all three numbers for stop bits with no success. One start bit is the only option. I have never seen a device that uses a different number of start bits. But serial is so screwed up I am sure there is someone out there that decided to use pi start bits! The bytestream can contain null characters, so I've found the Discard NUL property and switched it off. Since it's three wire communication- ground, rxd and txd, I've set handshaking to none. I've also disabled the termination character. Again on the offchance, I've tried sending an additional null byte at the end but it didn't help. Does anyone have any suggestions? It seems that your device is not responding because it probably hasn't gotten a complete message from the host. Check the documentation to see if there is a message terminator in the command string? Try a carriage return or lf as a terminator. Try EOT as a terminator. Make sure that the string has the correct check sum (sometimes protocols have the checksum include the string lenght tag, sometimes not). I would guess the problem is in carefully crafting the string since you have tried all the hardware options. Many systems have a way of snooping on serial port communications. I can give you a good way for Mac OS X.. But I am going out on a limb and guess that isn't your host system! :-) Others can suggest software for various winders versions. -Scott
What is the purpose of upgrading?
This contains very discouraging interview with Bill Gates. With a decidedly biased opinion piece, but certainly an educated discussion. http://billpalmer.net/com000144.html The full interview is http://www.jas.com/shame/shame/focus.html Obligatory LV content Do bugs get fixed between versions. Yesterday evening I looked over the bug list for LV from V 6 to 7.X. There are many long standing bugs, that just are not getting fixed! This is in 4 major upgrades of LV (some I know existed in LV 5.1!). A clear example is the thermometer units bug. Now it could be worked around by creating a VI, but it is still clear that it is a bug and fixing bugs in existing software has two drawbacks for a corporations. No glory and no income. References: Item 13, http://sthmac.magnet.fsu.edu/labview/buglist602.html Item 3, http://sthmac.magnet.fsu.edu/labview/buglist610.html Item 2, http://sthmac.magnet.fsu.edu/labview/buglist70.html There are several others as well. (waveform chart drawing, etc). FYI: these have all been reported to NI thru normal channels as well as beta test versions. -Scott (beating the dead horse of software reliability) -- Dr. Scott Hannahs, Head of User Research Instrumentation http://sthmac.magnet.fsu.edu National High Magnetic Field Laboratory, Florida State University 1800 E. Paul Dirac Dr., Tallahassee FL 32310, (850)644-0216/FAX 644-0534 Do not meddle in the affairs of sysadmins, for they are easy to annoy and have the root password.
Re: LabVIEW 7 Numeric Display precision - or Please Kill me now!
Jack, You wrote: Whos the darn fool at NI that set the default mode for numeric displays to Hide Trailing Zeros. ...And dont start that modify your Ini line I work on many different computers and cant go customizing the modifying every single one. I'm not sure it's any easier than modifying the .ini file, but at least it is unlikely that someone will modify it back if you put a customized control in you user.lib. -- EnWirementally, Paul F. Sullivan SULLutions (781)769-6869 when a single discipline is not enough visit http://www.SULLutions.com
RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!
Hi All I quickly put together asmall utility allowing to change selected controls on a VI panel. It allows enabling/disabling the "hide trailing zero" feature, and selection of the precision (it can be easily modified to do more if necessary). If someone is interested, contact me directly PJM Do you Yahoo!? Yahoo! Finance: Get your refund fast by filing online
Main VI stops when using control references.
Hi, This is my first question on this forum. I've created an engine simulation VI which uses control references. The main VI has a data acquistion loop and control references. The control references link to another VI that runs an equation loop which continuously calculates engine speed base on data collected in the main VI. The control references work but the main VI data is collected and updated only once when the main VI starts. If I remove the control references from the main VI the daq loop runs fine. Thanks in advance for your help. Mike
Re: i created boolean references in my main vi block diagram and...
Does the reference in the subVI say Bool Refnum on it or is it a Control Refnum? You also might try right clicking on the refnum in the subVI and selecting Include Data Type. This will turn on or off the strictly typed reference property. That's where I ran into problems. Rob
why I hate express vis.....
This is obviously my grumpy morning. And I have even had my prescribed dosage of coffee... I have been working with some software from a major instrumentation manufacturer. It is one of those VIs that tries to do everything with multiple tabbed front panels and lots of user interface controls and property nodes. Opening diagram extends over about 7 screens horizontally and 6 screens vertically (1600X1200 pixel screens). Ok, bad enough, now open the hierarchy window and I see a main VI and about 200 express nodes, that I can't open without converting to sub-VIs. There were 4 actual real non-system subVIs in the whole thing with main VI using about 12 Mb of disk space. Unfortunately the programmer had hardwired the path separation character so it failed on all platforms except the development one. The character was in 26 different places and each place was wired into somewhere between 1 and a dozen string building functions. -Scott
Re: Associate a type of file with a LV7 app
Dany Allard [EMAIL PROTECTED] wrote: Is it possible to associate a type of file, like *.txt with Notepad.exe, with a LV7 app. I know that LV7 have a new feature that permit to retrieve the command line used to call the app. But I dont think that would be useful for what I want to do. I want to Double-Click on my file and my LV7 app start automatically and the file is read and displayed. You will need the command line too! But first you need to go into the registry and create the according keys to associate your file ending with your application. Looking in the registry how things are done for instance for .vi files should give you enough of an idea. In the ROOT key you have a key .vi with its value set to some sort of name in this case LabVIEWInstrument. Under the same root group you get another group with that LabVIEWInstrument name and some groups and keys in there. The important one is shell/open/command whose value is the path to your executable and then %1 as parameter to tell the Windows shell to insert the file name the user has clicked on at this place. Then read the command line in your application and do what you need to do with it. Rolf Kalbermatter CIT Engineering Nederland BVtel: +31 (070) 415 9190 Treubstraat 7H fax: +31 (070) 415 9191 2288 EG Rijswijkhttp://www.citengineering.com Netherlands mailto:[EMAIL PROTECTED]
RE: Dual monitors for labview use
What is the difference in performance I could see between running two separate pci cards, one pci card and one agp card, or one dual-monitor card? -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Behalf Of Simon Whitaker Sent: Friday, February 13, 2004 2:59 AM To: [EMAIL PROTECTED] Subject: Re: Dual monitors for labview use Hi Scott, Can anyone tell me what graphic cards work well with respect to using dual monitors and labview? How does the graphic card handle moving the display data from one monitor to another. I was considering the ATI Radeon 9800XT. Great for gaming as well. I remember someone talking about having the diagram window on one machine and the panel window on another. No more fussing around with the windows to do your work faster. We use dual monitors here, using a variety of different graphics cards. I have an nVidia GeForce FX5200 (great card, but take care if you buy one because they do single- and dual-monitor varieties), my boss has a Matrox card of some description, and one of our machines just uses two separate single-display cards. That machine runs XP, which has dual-monitor support built in, so you can use two separate single-display cards instead of a single dual-display card. In all cases, LabVIEW works just fine with them. The software we develop uses a subVI to retain the screen position of VIs and put them back in the same place next time they're launched (using the VI's Windows.Bounds property), and that has no problem coping with the dual display setup. As someone else has noted, XP and Windows 2000 treat dual monitors as one large desktop, so software doesn't need to be aware of which monitor it's running on, it just runs somewhere on that large desktop and neither knows nor cares whether that happens to be on monitor 1 or monitor 2. The only noticable difference between the various solutions we use is the facilities offered by the driver software that comes with the video card(s). My nVidia card comes with an app called nView that allows you to do all sorts of fancy things with your dual monitors, such as enabling window spanning across desktops, choosing which monitor specific apps open on, setting different wallpapers for each monitor and so on. The machine that uses two separate cards and relies on the dual monitor support built into XP lacks these advanced features, although it's still perfectly usable. If you have an XP machine with a spare PCI slot, a spare PCI graphics card and a spare monitor, try plugging the card into the PC and giving it a go - won't cost you anything, and you'll see it first-hand. Try before you buy! :-) One other off topic question. Have you guys/gals been buying and using LCD monitors? Anyone running them for a long time? I keep seeing used monitors show up on ebay that have a few/many lcds burnt out or broken in the display. How hard and expensive are they to repair if this is the case? I can solder as I'm a EE by nature. I've got an LCD monitor as my 2nd display - a fairly cheap 15 (1024x768) model that I have pivoted through 90 degrees (another feature supported by the nVidia software), making it great for working on documents. I've had it for about 4 months and haven't had any problems so far with dead pixels. We've got another LCD display that we've had for about a year, again no problems with dead pixels. Not sure how feasible it is to repair a dead pixel. Note that even new LCD monitors may ship with some dead pixels, and most manufacturers specify a tolerance level for dead pixels below which they won't replace a panel. All the best, Simon Whitaker [EMAIL PROTECTED] Software developer, Tiab Ltd tel: +44 (0)1295 714046 fax: +44 (0)1295 712334 web: http://www.tiab.co.uk/
Re: tutorial to build CGI application
Hello Philippe, Here is a link to the CGI tutorial we currently have on our website: http://zone.ni.com/devzone/conceptd.nsf/webmain/E8A8D4B9497A1E1E86256802007B8E06?opendocument. It is not step by step, but I think it will help you. Also, if you need help with building your application into an executable, please look at the following knowledge base: http://digital.ni.com/public.nsf/websearch/17427B623FE6083F86256D1300705CFE?OpenDocument I hope this helps. Regards, Jyoti F. National Instruments Also, if you need help with
Re: Able to reset the board from MAX but not from my LABVIEW6.1 program
Jonnie 5 gave you the best advice regarding solving your problem - post your code. I just wanted to add that when you say Labview standard program, it doesn't mean an awful lot. There are several shipping examples of serial communication and they are different depending on which versioon of LabVIEW that you're using. Real old ones used the serpdrv and more recent ones use VISA. I don't remember any of the examples being setup to add a termination character to the write string. This is one of the most common mistakes people make when starting to program serial in LabVIEW.
NI reduced prices on some DAQ equipment
I was on NI's website and noticed that they are offering discounts on several of their popular DAQ devices... Dan Bookwalter N8DCJ Hi-Stat a Stoneridge Co. 345 South Mill St. Lexington, Ohio 44904 (419) 884-1219 Phone (419) 884-4172 Direct (419) 884-4195 Fax [EMAIL PROTECTED] _ This electronic mail transmission contains confidential information intended only for the person(s) named. Any use, distribution, copying or disclosure by any other person is strictly prohibited. If you received this transmission in error, please send an electronic mail message to [EMAIL PROTECTED]
Re: Dual monitors for labview use
I seem to have more limitations imposed on me by the monitor than the video card. According to the G550 specs, it has dual integrated RAMDACs, a 360MHz Primary and a 230 MHZ Secondary. It claims up to [EMAIL PROTECTED] on the main display, but my monitor can't go that fast. Likewise, the secondary should be able to go faster than 85Hz @1024x768, but once again the monitor can't. So maybe your own limitations aren't the video card? Just something to consider. John Howard Uwe Frenz 02/13/04 10:37AM John, you wrote: I am using a Matrox G550, and it works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED]) Thats about or little above the specs, AFAIK. The highres requires a pixel clock of 160 MHz. The video bandwidth of that chips is not far beyond 250 MHz, so the signal quality will not be very impressive. But it works - i use exactly the same resolution on a Nokia 445Xpro. But the second head used to have a much smaller video bandwidth. The producers tend to overstate their product capabilities, e.g. allowing resolutions that do not produce good signal quality - just to have impressive marketing labels. -- Mit freundlichen Grüßen! Greetings from Germany! -- Uwe Frenz ~ Dr. Uwe Frenz Entwicklung getemed Medizin- und Informationtechnik AG Oderstr. 59 D-14513 Teltow Tel. +49 3328 39 42 0 Fax +49 3328 39 42 99 [EMAIL PROTECTED] www.Getemed.de
Re: Filtering waveform?
LabVIEW 6.1 offers some few resampling options with the low level VIs. (Linear and Spline interpolation). See for example the VI attached to this reply: http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=137HOID=5065000500CE580100HTHREAD=88270UCATEGORY_0=_49_%24_6_UCATEGORY_S=0
RE: Serial comms
Hello Craig, I would like to recommend PortMon -- a freeware utility to monitor all serial (and parallel) port activity with myriad features. http://www.sysinternals.com/ntw2k/freeware/portmon.shtml I've found it to be an invaluable tool for troubleshooting serial comm issues.. in fact we joke about it as PortMon -- An Engineer's PokeMon! Hope you find it equally useful. Regards, Khalid Ansari - Forwarded by Khalid Ansari on 02/13/2004 11:29 AM - Craig Graham [EMAIL PROTECTED] Sent by: [EMAIL PROTECTED] 02/13/2004 06:30 AM To: [EMAIL PROTECTED] cc: Subject: Serial comms [clip] Since I have a laptop and 6036 card lying around, and I have a demo program for this reader, I think I'm going to end up doing a simple serial analyser and seeing what the differences are between my code and the demo. But that's not going to happen till next week and it will waste quite a bit of time so I thought I'd see if anyone had suggestions. [clip]
RE: Dual monitors for labview use
At 10:34 -0600 02/13/2004, Scott Serlin wrote: What is the difference in performance I could see between running two separate pci cards, one pci card and one agp card, or one dual-monitor card? Depends on what else you use the PCI bus for. It is slower performance for raw bit pushing than the AGP so if you have a smaller monitor use that one on the PCI bus. If you are also trying to acquire 60 MS/s of data, then I wouldn't recommend the PCI bus. It is a limited resource like everything else.
Re: Main VI stops when using control references.
Assuming you use the VI server to load the subvi dynamically, when loading the vi, the Wait to finish must be set to false, so your main loop does not wait for the subvi to finish. When the main loop is waiting for the subvi to finish, that case (loop) is halted. Joe
Using Tabs with Subvi's
I wish to have each page of a tab control be used with a subvi. One for graphs, one for data entry,etc and wish to have the tab page number start and stop the selected subvi. Up until now I have been using 'run on open' and a seperate stop button for each subvi. I tried creating dummy tab pages and local variables and it works but is clunky. The problem is that I can't read a tab page number selected from a main program while I'm in a subvi. Thanks, Jim W
Re: Labview PDA using RS232
I was looking for a solution to this question!
Re: Why does converting a SGL to SGL get rid of coersion dots?
The algorithm for determining resulting data types gets much more complicated when using uninitialized shift registers. The algorithm used by the binary arithmetic functions is slightly different than the algorithm used by the compound arithmetic function. It was determined that the algorithm used by the compound arithmetic function is more robust than that used by the binary arithmetic functions. Modification to the algorithm used by the binary arithmetic functions is being considered for future versions of LabVIEW.
Re: Problem with DAQMX digital output
Hi V.GRAF, The DAQ 6025E is a special card in terms of digital I/O. This is because it has two different semiconductor chips handling the digital. It has the DAQ-STC chip handling the first port of 8 bits (Port 0). It then has the 8255 chip handling the 3 other ports. In general, you should be able to use any of the DAQmx programs for digital with this card. However, you have to make sure that you are configuring the entire port at the beginning of your program. This is because if you reconfigure any of the 3 ports on the 8255 PPI chip, it will reset the port. So my suggestion is to not use line configuring but instead use port configuring. Here are a couple Knowledge Bases that discuss the issue. The NI PCI-6025E Digital Line Virtual Channels Return Error -10007 in Measurement Automation Explorer Test Panel http://digital.ni.com/public.nsf/websearch/77DBEDFCCA418A8DC1256C16007FA3B7?OpenDocument Why Do I Lose the Data on Ports 2, 3, and 4 of a PCI-6025E When Reconfiguring Port 0 in Visual Basic? http://digital.ni.com/public.nsf/websearch/33BE87B9752B7D7186256B1E0050A1B9?OpenDocument Hope that helps. Have a good day. Ron Applications Engineering National Instruments
Re: Sequence tunnel error
Ok guys here where I stand. I tried to pull the significant parts of the code out and put them in a new VI soo that i could post it to the board, but When i did so the new VI worked flawlessly. So i toyed around more removin parts of the sequence and different frames. I have narrowed it down to one specific area. Once the data was inside the sequence it goes to a subvi that is used for updating. This subvi output is the data is then fed to the hardware and fed to that tunnel thats giving me the trouble. Ok when the data leaves that subvi it goes to the output tunnel but it also goes to a sequence local to be used in a latter frame. This is the part thats causeing the change of data. If i remove that sequence local and instead use a local variable( i know another no no..) it works fine. Its as if the sequence local somehow interacts with the seqence tunnel. Im not sure if these data values are stored in the same memory spot or what but having that sequence local and the sequence tunnel have the same data value seems to screw everything up. For now im just going to use the local variable and be happy i have a work around. If i can figure out how to duplicate the problem ill post the code. I would post the entire VI but its got alot of hardware funcionality that im not sure what it would do to a system with out the hardware. Thanks for offering suggestions, they did lead me to a fix.
Re: subPanel: The coordinate system seems all scr... up? AND How...
The top-left corner of the subPanel uses the same coords as the top-left corner of your saved sub VI. Also the subPanel document size will be the same size as your running VI. I.E. the sub VI's should have the same placement / appearance as regular VIs, or exsiting in a subPanel. If you want to make sure they match up, then save your sub VIs with the origin at the top-left (0,0). One way to do this is to place a property node on your block diagram, right click the header to change the property class to VI and then select origin property, right click to change to write. Right click again to create a cluster (0,0). When you run your VI, the top-left will align perfectly with the two thick lines. Save your VI so that the VI will open at this coord (even in the subPanel). -David
NI SEC Invoke Loging Dialog
I=B4m using a NI SEC Invoke Loging Dialog.vi in my application but when I compile it in a stand alone .exe file, the NI SEC Invoke Loging Dialog doesn=B4t work.
Re: Some Problems with LabVIEW and Matlab DLLs
Hello spj. I have taken a look at your issue. It sounds like what is happening could be expected behavior. When you re-build your script into a DLL, as is mentioned in your post, it will re-register the COM object to Windows. Therefore, even though the names and locations are the same, the DLL (really like an .ocx file in this case) has been re-registered and thus will need to be re-linked to the LabVIEW Automation Open function each time. I hope this helps! Colin C. Applications Engineer
Re: How to compile wavread on MatLab Script?
Janet, Remove the quotes around blah. Just put in your MATLAB Script Node f=wavread(blah) With the quotes, MATLAB was trying to read the string 'blah', and not the blah input, which is 'annano.wav' Zvezdana S. National Instruments
Re: 34970A VI?
If you're talking about an HP/Agilant, there are vi's included with the driver: http://zone.ni.com/idnet97.nsf/9b2b33e1993d877786256436006ec498/e4c553c139f78062862568ab005fba3c?OpenDocument See also: http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=50650008001786UCATEGORY_0=_49_%24_6_UCATEGORY_S=0USEARCHCONTEXT_QUESTION_0=34970AUSEARCHCONTEXT_QUESTION_S=0 Or just search the discussion forums for 34970A...
Re: How do you reset the encoder count with LV?
STK, Yes Rest Encoder Count is for resetting a unmapped encoder. For an encoder that is mapped use Reset Position This will change the value of the encoder that is mapped to an axis. A. Talley Application Engineer National Instruments
Re: Unable to lock memory chunk, unable to allocate memory
Hello Matute, Thank you for contacting National Instruments. Try looking at the discussion thread linked below. It concerns a problem similar to yours. http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RNAME=ViewQuestionHOID=5065000800BEA7ECategory=LabVIEW.LabVIEW+General Let me know if this does not help. Sean C. Applications Engineer National Instruments
Re: Write hex code to pic with serial port, any example?
Hello Jose, Thank you for contacting National Instruments. The following link leads to a discussion thread about hex to ASCII serial communication. http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RNAME=ViewQuestionHOID=5065000800F544ECategory=Instrument+Connectivity.Serial+Interfaces Please reply to this post if you have some more specific questions. I hope this helps, Sean C. Applications Engineer National Instruments
Re: Sequence tunnel error
One of the problems with the sequence structure is that data doesn't leave any of the frames until the last frame is done. I suspect that that is what you are seeing if the data is modifed in a later frame.
Re: how does the interface to ADO drivers work?
I'm using the ActiveX interface. Mike...
low level device drivers
OK, I think I am in over my head. :) I am considering using a PCI card (Sabtech NTDS interface card for those interested), and it doesn't seem to have a nice simple DLL like I am use to. Instead it has a kernel mode Windows 2000 device driver, and I am expected to use ...normal Windows I/O system calls... to talk to it. The driver is written to be part of the Serial Port driver class, BUT it can not be treated simply as a serial port since ...there are several restrictions on the I/O calls..., and ...it has its own set of DeviceIoControl functions. I know so little about Windows device drivers that I don't really even understand the paragraph I just wrote. Does this mean I can't talk to the card directly from LabVIEW? If I can, is it difficult for someone who uses DLL's fairly often? Can I use the Win32 API to talk to it? Does anyone have any suggestions on how to approach this? There is another card I am considering from a different company that does have a DLL (with a rather obscure command set), which I could use instead of this one. Or maybe I could get one of our .NET programers to write me something simple. Suggestions anyone? Thanks for any nuggets of wisdom! John Howard
Re: How can i display the signal out of a FFT block to a graph?
Hello St. Augustine, You should not need to do any conversion from the output of the FFT VIs to the waveform graph. Make sure, however, that you are wiring into a Waveform Graph.vi and not the Waveform Chart.vi. Regards, Jyoti F. National Instruments
Re: how can I create a simulated hardware configuration in MAX Measurement Automation Explorer
There really isn't anyway to fake having a DAQ board, and w/o a DAQ board, you're going to get errors when trying to communicate to it. You can either: 1) Remove the hardware commands and use simulated dsta. http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800A162UCATEGORY_0=_49_%24_6_UCATEGORY_S=0 2) Use an alternative DAQ such as a serial/parallel port or even a sound card http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800C7B2UCATEGORY_0=_49_%24_6_UCATEGORY_S=0USEARCHCONTEXT_QUESTION_0=simulatedUSEARCHCONTEXT_QUESTION_S=0 3) Connect to the PC w/ the DAQ remotely to test the program (Remote Data Acquisition or RDA). http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800E76EUCATEGORY_0=_30_%24_12_UCATEGORY_S=0USEARCHCONTEXT_TIER_0=0USEARCHCONTEXT_TIER_S=0USEARCHCONTEXT_QUESTION_0=simulatedUSEARCHCONTEXT_QUESTION_S=0
Re: Main VI stops when using control references.
the Engine Momentum Subvi should finish execution with each iteration of the loop in the DAQ.vi Therefore , connect a constant Boolean set to True to the condiotional terminal of the while Loop of the Engine Momentum Subvi (instead of the stop value from the main vi)
Re: Main VI stops when using control references.
I tried that and it works but the zero outside the loop which resets the shift register to zero each time the loop starts must be removed so the engine accelerates. I just need to figure out how to reset the speed to zero after I've stopped and restarted the main vi without have to close it and reopen. Thanks, Mike
Re: Read Trajectory Data Function and 7344 card
Chris, The difference is the hardware 3ms is a 7350 and 7344 is the 10 ms. The diference is which card you by. A Talley Applications Engineering National Instruments
RE: Dual monitors for labview use
Can this feature, two independent desktops be implemented in win2k or winXP? Also, regarding the potential invisible area on the screen, can't you just go to a small resolution on the larger screen so that the two screen match? I have to commend this list as I would never have thought about a dual-monitor setup if one poster (who I have forgot-sorry) had not mentioned it. I was on a website recently and they had a rack to set up multiple LCD screens on one monitor rack. I wonder what the video card would look like if trying to drive 3 or more monitors. What's out there just for grins? How expensive? I could have my labview diagram on one screen, panel on the other, dvd video playing on a third, etc. I just need to install more eyes into my head to view all of this. 8-) On Win9x however there will be just two independend desktops -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Behalf Of Uwe Frenz Sent: Friday, February 13, 2004 1:50 AM To: Scott Serlin Cc: LV-Info, list Subject: Re: Dual monitors for labview use Scott, you asked on Thu, 12 Feb 2004 13:26:27 -0600: Can anyone tell me what graphic cards work well with respect to using dual monitors and labview? How does the graphic card handle moving the display data from one monitor to another. I was considering the ATI Radeon 9800XT. Great for gaming as well. I remember someone talking about having the diagram window on one machine and the panel window on another. No more fussing around with the windows to do your work faster. Well, here at getemed most of the devellopers have now a dual monitor system. Most of us have matrox G550s, but almost all dual monitor cards should work well too. I had a project where the customer even wanted three large monitors. So I added two graphics cards, an ATI and a Matrox dualhead into the machine. All worked well, allthough (because of the 3 monitors) I had to use Win98 at that time. Here we come to a big caution: Win NT 2k (no knowledge on XP yet) handle multiple monitors as two windows to a large and unique desktop, whereas Win9x used to have multiple independend desktops that just had to share one side with each other. This might be important, because almost all dualhead cards I know of have an excellent head as primary head, supporting video bandwidth of up to 250 MHz (and resolutions of up to 1600*1200). The secondary head is propably less capabal, supportimg only smaller resolutions. This may result in some area on the desktop that is not displayed on any of the screens in W2k. Example: One screen with 1600*1200 left and another with 1280*1024 right aligned at bottom. The desktop will be (1600+1280)*1200 and there will be an area of (1600, 1199) to (2879, 1024) that will be invisibel. On Win9x however there will be just two independend desktops. Also, how about those kvm monitor switches? Anyone used those in conjunction with the dual monitor graphic cards? Is going with an extra PCI board as well as the ATI Radeon 9800XT AGP board the better route than running both monitors off of the same card? Thoughts? Comments? Monitor switches have to support the high video bandwidth in order not to damage signal quality. Try it out for the equipment and resolution you are trying to use. One other off topic question. Have you guys/gals been buying and using LCD monitors? Anyone running them for a long time? I keep seeing used monitors show up on ebay that have a few/many lcds burnt out or broken in the display. How hard and expensive are they to repair if this is the case? I can solder as I'm a EE by nature. AFAIK this is an internal damage that can't be repaired and never ever with a 'burning iron'. Nothing to say about your 'natural' solder art crafting. (Hope this is got as that joke it is intended to be. Joking in a foreigh language is not so easy.) The LCD producers have even defined a level of defective pixels that are considered as OK according to the production standards. Otherwise they would get too few panels being OK out of their production lines and those would be too expensive for mass market. Greetings from Germany! -- Uwe Frenz ~ Dr. Uwe Frenz Entwicklung getemed Medizin- und Informationtechnik AG Oderstr. 59 D-14513 Teltow Tel. +49 3328 39 42 0 Fax +49 3328 39 42 99 [EMAIL PROTECTED] WWW.Getemed.de
Re: subPanel: The coordinate system seems all scr... up? AND How...
Thanks! - Martin
Serial Port
Hi I have done data acquisition using NI DAQ card for one of my projects. Right now, have a small project in which data needs to be acquired from a serial port. The system has a temperature controller unit which sends data to the serial port and needs to be collected using LabView. How different is it to configure and acquire data using serial port rather than NIDAQ card. Could anyone help me out in this? Also, this is going to be done using LabView 7 - Evaluation Version. I guess there are no restrictions while we use evaluation version. I hope all the features of a full development system could be used. Thanks for the help, Shankar
scaling graph to frequency?
i have succesfully read in data from a digital input and managed to store in a text file. This data appears to be most likely an ASCII format. Are there ASCII to number converters? I would also like to now use this data and perform analysis on them.For example using the FFT etc. I have managed to read the data from the same file and perform analysis on it.However the FFT must represents a signal in the frequency domain and the scale remains in time. Im not sure how to scale my X axis to represent this? Can you please help!
RE: Dual monitors for labview use
On Fri, 13 Feb 2004, Scott Serlin wrote: What is the difference in performance I could see between running two separate pci cards, one pci card and one agp card, or one dual-monitor card? most likely, zero difference if you by a recent agp dual head card. possibly fewer driver issues with a single card as well. -k
Can we acquire the Analog input and Digital Input at
Can we acquire the Analog input and Digital Input at the same time? My sensor data include Analog and digital signals.
RS232 Displays for LVRT
Hi all, Does anyone know of any display (preferably touchscreen, but not necessarily) that can be controlled over RS232 interface (rather than a video board or chipset)? I need to connect a local display to a system running LabVIEW RT, and it looks like RS232 is the only option that I have. I don't need a VGA resolution - smaller resolution, black-and-white LCD will be fine. I know there was at least one company that had products like that - it had a booth at NIWeek a couple years ago - but I can't recall its name. TIA, Sergey Liberman [EMAIL PROTECTED] Solidus Integration http://solidusintegration.com