Re: Dual monitors for labview use

2004-02-13 Thread Uwe Frenz
Scott,

you asked on Thu, 12 Feb 2004 13:26:27 -0600:
 Can anyone tell me what graphic cards work well with respect to using
 dual monitors and labview?  How does the graphic card handle moving the
 display data from one monitor to another. I was considering the ATI
 Radeon 9800XT.  Great for gaming as well. I remember someone talking
 about having the diagram window on one machine and the panel window on
 another.  No more fussing around with the windows to do your work
 faster.  
Well, here at getemed most of the devellopers have now a dual monitor
system. Most of us have matrox G550s, but almost all dual monitor cards
should work well too.

I had a project where the customer even wanted three large monitors. So I
added two graphics cards, an ATI and a Matrox dualhead into the machine.
All worked well, allthough (because of the 3 monitors) I had to use Win98
at that time.

Here we come to a big caution:
Win NT  2k (no knowledge on XP yet) handle multiple monitors as two
windows to a large and unique desktop, whereas Win9x used to have multiple
independend desktops that just had to share one side with each other. This
might be important, because almost all dualhead cards I know of have an
excellent head as primary head, supporting  video bandwidth of up to 250
MHz (and resolutions of up to 1600*1200). The secondary head is propably
less capabal, supportimg only smaller resolutions. This may result in some
area on the desktop that is not displayed on any of the screens in W2k.
Example: One screen with 1600*1200 left and another with 1280*1024 right
aligned at bottom. The desktop will be (1600+1280)*1200 and there will be
an area of (1600, 1199) to (2879, 1024) that will be invisibel.
On Win9x however there will be just two independend desktops.

 Also, how about those kvm monitor switches?  Anyone used those
 in conjunction with the dual monitor graphic cards?  Is going with an
 extra PCI board as well as the ATI Radeon 9800XT AGP board the better
 route than running both monitors off of the same card?  Thoughts?
 Comments?  
Monitor switches have to support the high video bandwidth in order not to
damage signal quality. Try it out for the equipment and resolution you are
trying to use. 

 One other off topic question.  Have you guys/gals been buying
 and using LCD monitors?  Anyone running them for a long time?  I keep
 seeing used monitors show up on ebay that have a few/many lcds burnt out
 or broken in the display.  How hard and expensive are they to repair if
 this is the case? I can solder as I'm a EE by nature.
AFAIK this is an internal damage that can't be repaired and never ever with
a 'burning iron'. 
Nothing to say about your 'natural' solder art crafting.
(Hope this is got as that joke it is intended to be. Joking in a foreigh
language is not so easy.)

The LCD producers have even defined a level of defective pixels that are
considered as OK according to the production standards. Otherwise they
would get too few panels being OK out of their production lines and those
would be too expensive for mass market.

Greetings from Germany!
-- 
Uwe Frenz


~
Dr. Uwe Frenz
Entwicklung
getemed Medizin- und Informationtechnik AG
Oderstr. 59
D-14513 Teltow

Tel.  +49 3328 39 42 0
Fax   +49 3328 39 42 99
[EMAIL PROTECTED]
WWW.Getemed.de




Re: Vi converted to LabView 7.0 Error

2004-02-13 Thread waldemar.hersacher
Hi Dr. Imad,

another workaround is described a
href=http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800939BUCATEGORY_0=_49_%24_6_UCATEGORY_S=0;here/a.
Search for my answer. This is a conversion bug in LV 7.0 and the
workaround is first load it into LV 7.0 then save it into the new
location and then make your changes.

Waldemar



Re: Dual monitors for labview use

2004-02-13 Thread Simon Whitaker
Hi Scott,

 Can anyone tell me what graphic cards work well with respect to using
 dual monitors and labview?  How does the graphic card handle moving the
 display data from one monitor to another. I was considering the ATI
 Radeon 9800XT.  Great for gaming as well. I remember someone talking
 about having the diagram window on one machine and the panel window on
 another.  No more fussing around with the windows to do your work
 faster.

We use dual monitors here, using a variety of different graphics cards.
I have an nVidia GeForce FX5200 (great card, but take care if you buy
one because they do single- and dual-monitor varieties), my boss has a
Matrox card of some description, and one of our machines just uses two
separate single-display cards. That machine runs XP, which has
dual-monitor support built in, so you can use two separate
single-display cards instead of a single dual-display card.

In all cases, LabVIEW works just fine with them. The software we develop
uses a subVI to retain the screen position of VIs and put them back in
the same place next time they're launched (using the VI's Windows.Bounds
property), and that has no problem coping with the dual display setup.
As someone else has noted, XP and Windows 2000 treat dual monitors as
one large desktop, so software doesn't need to be aware of which monitor
it's running on, it just runs somewhere on that large desktop and
neither knows nor cares whether that happens to be on monitor 1 or
monitor 2.

The only noticable difference between the various solutions we use is
the facilities offered by the driver software that comes with the video
card(s). My nVidia card comes with an app called nView that allows you
to do all sorts of fancy things with your dual monitors, such as
enabling window spanning across desktops, choosing which monitor
specific apps open on, setting different wallpapers for each monitor and
so on. The machine that uses two separate cards and relies on the dual
monitor support built into XP lacks these advanced features, although
it's still perfectly usable.

If you have an XP machine with a spare PCI slot, a spare PCI graphics
card and a spare monitor, try plugging the card into the PC and giving
it a go - won't cost you anything, and you'll see it first-hand. Try
before you buy! :-)

 One other off topic question.  Have you guys/gals been buying
 and using LCD monitors?  Anyone running them for a long time?  I keep
 seeing used monitors show up on ebay that have a few/many lcds burnt out
 or broken in the display.  How hard and expensive are they to repair if
 this is the case? I can solder as I'm a EE by nature.

I've got an LCD monitor as my 2nd display - a fairly cheap 15 (1024x768)
model that I have pivoted through 90 degrees (another feature supported
by the nVidia software), making it great for working on documents. I've
had it for about 4 months and haven't had any problems so far with dead
pixels. We've got another LCD display that we've had for about a year,
again no problems with dead pixels. Not sure how feasible it is to
repair a dead pixel. Note that even new LCD monitors may ship with some
dead pixels, and most manufacturers specify a tolerance level for dead
pixels below which they won't replace a panel.

All the best,


Simon Whitaker [EMAIL PROTECTED]
Software developer, Tiab Ltd
tel: +44 (0)1295 714046
fax: +44 (0)1295 712334
web: http://www.tiab.co.uk/




motion control feedback (analog) PID

2004-02-13 Thread wayne galbraith
Hi there LV'ers

I have a few motion control problems that are, to quote ' doin my head in'.

I have a system that uses both encoder and analog feedback at varying times 
during the program operation.  This all seems to work fine with the system 
configuring and re-configuring the analog/encoder etc what ever is required.

My system is then to perform two task under analog control, one is to 
simply drive to an analog feedback value and hold, the other is to cycle 
between two analog feedback values.

The main part of the problem that I have is the setting of PID parameters 
using the MAX autotune facility.  This autotune feature works fine with 
encoder feedback with values in the range of 10 .. 250 for the 3 main PID 
parameters.  When this same task is performed with analog feedback the 
machine drives at high velocity and comes in with values of 32767 for the 
PID parameters - clearly an error.

If we try to set the PID parameters for analog feedback manually we come up 
with a few values in the 10 .. 100's range for the parameters and the motor 
sort of behaves itself except when we try a simple cycle between two values 
in the MAX 1-D interactive screen.  The one direction move say to 200 
(whatever's) works fine, a cyclic move from 200 to 600 will cycle the motor 
very hard and not actually stop on the desired values etc.

These problems seem as though they are all PID problems and I just can't 
work out why I can't auto tune on the analog feedback.  The transducer we 
are using is a load cell, although I have changed this to an LVDT just to 
ensure there wasn't a noise problem on the signal.

Thanks for any help

wayne

--
Wayne Galbraith
School of Civil and Resource Engineering
The University of Western Australia
35 Stirling Highway
Crawley 6009
Tel - (08) 9380 3788
email - [EMAIL PROTECTED]
--





Re: network oscilloscope

2004-02-13 Thread Stijn
Matthew,
thanks for your answer. The Scope Chart Mode is very interesting..
I actually wanted to know how to implement the way a DSO works in LV
(setting trigger level and timebase etc.), but this isn't necessary
anymore since I decided to use my dsp to calculate the data to be
plot, hence acting as a DSO. Then it sends this data to LV, so LV
should only display it, and I'm probabely going to use the Scope Chart
for that.

-Stijn



I need that Programm for LV 5.0

2004-02-13 Thread germannubie
thank you jonnie 5 but there are only programs for LV 6.0
can I download programs for version 5.0 or a LV-Update



Re: How to compile wavread on MatLab Script?

2004-02-13 Thread Shan
When you say it will not compile, do you mean you've got a broken
arrow or you get an error?  If broken arrow, what does the error list
say the problem is?  If you're getting an error out of the Matlab
Script structure, what is it?

I've tried this on LV 6.1 and Matlab 6.5 and it works fine for me.
Post your code  we'll try to help.

Tim



Re: How to compile wavread on MatLab Script?

2004-02-13 Thread Shan
Also, check out this link - don't know if it addresses your problem.

http://digital.ni.com/public.nsf/websearch/4475BC3CEB062C9586256D750058F14B?OpenDocument



Re: DAQ+storage must happen independent of other screens OR user interaction

2004-02-13 Thread waldemar.hersacher
There are some examples how to use named queue.

LV 6.0.2: Search Examples  Advanced Examples  Execution control
Examples  Synchronisation examples  Named Queue

LV 6.1  LV 7.0: Find Examples  Optimizing Applications 
Synchronizing Tasks  Queue Message Loging.vi

In this example they use multpile sender and one receiver, but they
show using a named queue to communicate between parallel running VIs.

In the same example categories are also examples for notifiers who
demonstrate how to send data from one sender to multiple receivers.

I hope this will help you but feel free to ask again.

Waldemar



Re: Optimizing network throughput over Ethernet with TCP/IP

2004-02-13 Thread greenkiwi
Ethernet packets are range from 46-1500bytes, or there abouts, so
anything is being broken up.  Well, anything over 1500 bytes.

I'll give the even multiples of 32768 a try and see how that goes.

(Well, there seems to be support for frame sizes up to 9000 bytes in
Gigabit Ethernet.)

I'm going to have to experiment a bit more to see whether the delays
are caused.  I may also try to create an app in C or find a test app
that will pump data through, to test out LV's receive limitations.

Adam



Re: Why does converting a SGL to SGL get rid of coersion dots?

2004-02-13 Thread Ben
Well changing the setting and re-opening LV is not enough to fix this
issue.

Since the wire has to be dis-connected and then re-connected to get
rid of the coercion, this has to be an issue with what LV is doing at
edit time.

So...


Ben



RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!

2004-02-13 Thread Lewis Drake
Who’s the darn fool at NI that set the default mode for numeric displays to
“Hide Trailing Zero’s”.

Try setting the displays to digits of precision instead of significant
digits

The other annoyance (for DBL controls and indicators) is that they set the
default formatting to automatic with 6 significant digits.  How many times
do you want to see 6 digits after each number?  I'd like to see it set to
floating point with 2 digits of precision.

I suggest to NI that a method be included that will allow the LabVIEW
programmer to set these defaults in LabVIEW for all controls and indicators
globally.







i created boolean references in my main vi block diagram and...

2004-02-13 Thread Mudda
i created boolean references in my main vi block diagram and copied
them to my sub vi front panel. when wire my reference in my main vi to
one the input node of the sub vi the wire is broken. the error says
its a class conflict why?

i created boolean references in my main vi block diagram and copied
them to my sub vi front panel. when wire my reference in my main vi to
one the input node of the sub vi the wire is broken. the error says
its a class conflict why?



Re: Clusters: How to remove the 3D shaded background?

2004-02-13 Thread Ben
Well here is a golden nugget hidden at the bottom of a pile of fluff!

Transparent Front Panle Backgrounds!

I wounder what I could do with that?

Ben



Re: Dual monitors for labview use

2004-02-13 Thread Howard, John
 Uwe Frenz 02/13/04 02:49AM 
...but almost all dual monitor cards should work well too. ...

This may be true in general, although I saw one dual monitor card do nasty
things.  The brand name was something like Apien or Apex or something like
that.  (does that sound familiar to anyone?)  It was unbelievable how badly
this card performed.  Certain everyday tasks would only work on one monitor,
and there were other problems I forget now.  I am using a Matrox G550, and it
works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED])

John Howard




Re: divider, advice needed, please

2004-02-13 Thread Hamid_Yazdi

Hello guys;

Firstly let me thank a few people who answered my question, regarding
dividing a signal freq. 

The funny thing is that the whole concept looks so easy and so trivial, but IT DOES NOT WORK!

All the people who answered me told me to use Down Counter or Divider Config.Vi

I can swear to God that I have done this but it did not work and that is why 
I asked the list for some help. After all I don't think I am that daft! 

But I can tell now that I am stuck now.

So, the question is Has any body used the Down Counter or Divider Config.Vi, successfully?
Do you know where can I get a working example?

I am going to ask NI because this looks to be a question for them

Thanks again


+
Dr. Hamid R. Yazdi
Federal Mogul
Manufacturing technology
3935 Research park drive
Ann Arbor, MI 48108

Tel: 734 222 4108


Re: Problem with DAQMX digital output

2004-02-13 Thread BJD1613
Are you using the daqmx functions?  If so please post the code.  Or
you can try using the Daq assistant.  This will step you through the
process of configuring a digital output.  hope this helps.



Re: How can i display the signal out of a FFT block to a graph?

2004-02-13 Thread JFT'
Hi,  For the Y-axis, it depends of the input data binaries.  I mean,
if these come from a voltage signal or a collection of data or a sound
signal, etc.  It's necessary to know the factor conversion the voltage
values, for example...
 And for the X-axis, that represents the axis of frequencies, and its
posible to reconstruct until the nyquist order of the harmonic
frequency (FS/2).Also it's important, the lenght of the input
data, it's a power of 2, since algorithm FFT gives a fast and optimal
result.
  Then connect the output to a Waveform Graph Display, on the front
panel.
I hope this help.
Bye..



Re: Clusters: How to remove the 3D shaded background?

2004-02-13 Thread chutla
Don't you just hate it when that happens?  :)

Eric



Re: Filtering waveform?

2004-02-13 Thread timohe
How can i do re-sampling in LabVIEW 6.1?



RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!

2004-02-13 Thread Junginger, Johann

Jack,

I have started playing with LabVIEW 7 recently and this is already starting
to drive me batshit. My high school science teachers would dock marks for
failing to properly account for significant figures. 

To a physicist (and I assume for other technically inclined professionals)
0, 0.00 and 0.00 are three *different* numbers. And I haven't even
specified my uncertainties... :)

Excel also has this annoying default behaviour.

Sincerely,
Johann Junginger. 

-Original Message-
From: Jack Hamilton [mailto:[EMAIL PROTECTED] 
Sent: Wednesday, February 11, 2004 12:50
To: LabVIEW -Info
Subject: LabVIEW 7 Numeric Display precision - or Please Kill me now!


Who's the darn fool at NI that set the default mode for numeric displays to
Hide Trailing Zero's.

I am now getting carpal tunnel just correcting this every-single-time I drop
a numeric display on the front panel.

Why is God name would I declare a single or double precision number -  and
NOT want to see the values to the right of the decimal? Nearly without
exception clients want to see the darn decimal places.

And don't start that modify your Ini line - I work on many different
computers and can't go customizing the modifying every single one.

Does this seem like a rant - Yup ;- Fed by the 100's time I've had to
change the display configuration - I can do it now with my eyes closed.

This is like may SUV, the back door swings the wrong way - now I have to
stand in traffic to put stuff in the back - gee seemed like a good idea in
the development lab...

Jack Hamilton
Hamilton Design
[EMAIL PROTECTED]
www.Labuseful.com
714-839-6375 Office




Re: Dual monitors for labview use

2004-02-13 Thread Uwe Frenz
John,

you wrote:
 I am using a Matrox G550, and it
 works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED])
Thats about or little above the specs, AFAIK.
The highres requires a pixel clock of 160 MHz. The video bandwidth of that
chips is not far beyond 250 MHz, so the signal quality will not be very
impressive. But it works - i use exactly the same resolution on a Nokia
445Xpro.
But the second head used to have a much smaller video bandwidth. The
producers tend to overstate their product capabilities, e.g. allowing
resolutions that do not produce good signal quality - just to have
impressive marketing labels.

-- 
Mit freundlichen Grüßen!
Greetings from Germany!
-- 
Uwe Frenz


~
Dr. Uwe Frenz
Entwicklung
getemed Medizin- und Informationtechnik AG
Oderstr. 59
D-14513 Teltow

Tel.  +49 3328 39 42 0
Fax   +49 3328 39 42 99
[EMAIL PROTECTED]
WWW.Getemed.de




conflicts between property nodes

2004-02-13 Thread anne.M
I am using property nodes on image display to set the position center
( X and Y ) and the zoom (X and Y). When using only position center
property nodes everything works fine , but if i set the zoom value
with property nodes the position center doesn't work any more, the
image stays in the default position.
Is there any known reason for this behavior?
I am using Laview 7 under windows 2000 professionnel, and i am
acquiring an image with a PCI 1408 card.
Thanks,



Frequency spectrum of IFFT output

2004-02-13 Thread wes saunders
Output of IFFT block is summation of N orthogonal sinusoids. These
sinusoids each have different frequency and the lowest is 0 (DC). In
reality output is array of complex numbers which represent complex
weight for corresponding sinusoidal basis function. I neeed to know
how to present output signal in frequency domain spectrum. I'm
confused with complex numbers. Anybody knows solution to this problem?



Re: Problem with array processing !!!

2004-02-13 Thread Mads
It's due to the fact that you are displaying all those 100 000 values
and updating them in the indicator on every iteration.

If you move the indicator outside of the loop so that it will only be
updated when the processing is finished things will run fast.



Re: Problem with DAQMX digital output

2004-02-13 Thread Jonnie 5
Yes, the 6035 is supported (unless you're running win 95... or 3.1...
you are on windows, right?)
See:
http://www.ni.com/support/daq/versions.htm

The troubleshooter reffed above is at:
http://www.ni.com/support/install/



Re: unable to read the frequency from this output sine wave of amplitude VS time

2004-02-13 Thread JFT'
I don't really see the utility of this VI, however, the variable scan
rate, of the LJstream graph processing, it's an arbitrary value,
why?...
 What is the idea with these VIs?, tell me more, please.

Bye.



Re: mouse up event on a slide

2004-02-13 Thread tbob
Sorry, I misunderstood your problem.  I was able to re-create your
problem.  This could be considered a bug, moving off slider before
letting mouse up should return the pointer to its original value, but
it does not.  So now your event doesn't fire because it never saw a
mouse up event on the slider.  The only fix I could see is to set the
event trigger to Value Change.  Yes you will have an event for each
incremental change.  Don't know of a good solution.



Re: Serial comms

2004-02-13 Thread Scott Hannahs
At 12:30 + 02/13/2004, Craig Graham wrote:
The next is in the other port settings. The manual specifies 1 start bit, 8
data bits, 1 stop bit, no parity. The VISA serial stuff in Labview 7, which
I've only just started playing with, has properties for stop bits, data bits
and parity which I've set, but I can find no option for start bits. Is this
a problem? On the offchance, I've tried all three numbers for stop bits
with no success.
One start bit is the only option.  I have never seen a device that uses a different 
number of start bits.  But serial is so screwed up I am sure there is someone out 
there that decided to use pi start bits!


The bytestream can contain null characters, so I've found the Discard NUL
property and switched it off. Since it's three wire communication- ground,
rxd and txd, I've set handshaking to none. I've also disabled the
termination character.

Again on the offchance, I've tried sending an additional null byte at the
end but it didn't help.

Does anyone have any suggestions?

It seems that your device is not responding because it probably hasn't gotten a 
complete message from the host.  Check the documentation to see if there is a 
message terminator in the command string?  Try a carriage return or lf as a 
terminator.  Try EOT as a terminator.  Make sure that the string has the correct check 
sum (sometimes protocols have the checksum include the string lenght tag, sometimes 
not).  I would guess the problem is in carefully crafting the string since you have 
tried all the hardware options.

Many systems have a way of snooping on serial port communications.  I can give you a 
good way for Mac OS X..  But I am going out on a limb and guess that isn't your 
host system!  :-)  Others can suggest software for various winders versions.

-Scott




What is the purpose of upgrading?

2004-02-13 Thread Scott Hannahs

This contains very discouraging interview with Bill Gates.  With a decidedly biased 
opinion piece, but certainly an educated discussion.
http://billpalmer.net/com000144.html

The full interview is http://www.jas.com/shame/shame/focus.html

Obligatory LV content  Do bugs get fixed between versions.  Yesterday evening I 
looked over the bug list for LV from V 6 to 7.X.  There are many long standing bugs, 
that just are not getting fixed!   This is in 4 major upgrades of LV (some I know 
existed in LV 5.1!).

A clear example is the thermometer units bug.  Now it could be worked around by 
creating a VI, but it is still clear that it is a bug and fixing bugs in existing 
software has two drawbacks for a corporations.  No glory and no income.

References:
Item 13, http://sthmac.magnet.fsu.edu/labview/buglist602.html
Item 3, http://sthmac.magnet.fsu.edu/labview/buglist610.html
Item 2, http://sthmac.magnet.fsu.edu/labview/buglist70.html

There are several others as well.  (waveform chart drawing, etc).  FYI: these have all 
been reported to NI thru normal channels as well as beta test versions.

-Scott (beating the dead horse of software reliability)

-- 
 Dr. Scott Hannahs, Head of User Research Instrumentation
 http://sthmac.magnet.fsu.edu
 National High Magnetic Field Laboratory, Florida State University
 1800 E. Paul Dirac Dr., Tallahassee FL 32310, (850)644-0216/FAX 644-0534

  Do not meddle in the affairs of sysadmins,
   for they are easy to annoy and have the root password.




Re: LabVIEW 7 Numeric Display precision - or Please Kill me now!

2004-02-13 Thread Paul F. Sullivan
Jack,

You wrote:

Who’s the darn fool at NI that set the default mode for numeric displays to
“Hide Trailing Zero’s”.
...And don’t start that “modify your Ini” line – I work on many different
computers and can’t go customizing the modifying every single one.
I'm not sure it's any easier than modifying the 
.ini file, but at least it is unlikely that 
someone will modify it back if you put a 
customized control in you user.lib.

--
EnWirementally,
Paul F. Sullivan


SULLutions  (781)769-6869
when a single discipline is not enough
	visit http://www.SULLutions.com






RE: LabVIEW 7 Numeric Display precision - or Please Kill me now!

2004-02-13 Thread PJ M
Hi All

I quickly put together asmall utility allowing to change selected controls on a VI panel.
It allows enabling/disabling the "hide trailing zero" feature, and selection of the precision (it can be easily modified to do more if necessary).
If someone is interested, contact me directly

PJM

Do you Yahoo!?
Yahoo! Finance: Get your refund fast by filing online

Main VI stops when using control references.

2004-02-13 Thread mkdcs
Hi, This is my first question on this forum.  I've created an engine
simulation VI which uses control references. The main VI has a data
acquistion loop and control references. The control references link to
another VI that runs an equation loop which continuously calculates
engine speed base on data collected in the main VI.  The control
references work but the main VI data is collected and updated only
once when the main VI starts.  If I remove the control references from
the main VI the daq loop runs fine.  Thanks in advance for your help.

Mike



Re: i created boolean references in my main vi block diagram and...

2004-02-13 Thread Robert Cole
Does the reference in the subVI say Bool Refnum on it or is it a
Control Refnum? You also might try right clicking on the refnum in
the subVI and selecting Include Data Type. This will turn on or off
the strictly typed reference property. That's where I ran into
problems.

 Rob



why I hate express vis.....

2004-02-13 Thread Scott Hannahs
This is obviously my grumpy morning.  And I have even had my prescribed dosage of 
coffee...

I have been working with some software from a major instrumentation manufacturer.  It 
is one of those VIs that tries to do everything with multiple tabbed front panels and 
lots of user interface controls and property nodes.

Opening diagram extends over about 7 screens horizontally and 6 screens vertically 
(1600X1200 pixel screens).  Ok, bad enough, now open the hierarchy window and I see a 
main VI and about 200 express nodes, that I can't open without converting to sub-VIs.  
There were 4 actual real non-system subVIs in the whole thing with main VI using about 
12 Mb of disk space.

Unfortunately the programmer had hardwired the path separation character so it failed 
on all platforms except the development one.  The character was in 26 different places 
and each place was wired into somewhere between 1 and a dozen string building 
functions.

-Scott




Re: Associate a type of file with a LV7 app

2004-02-13 Thread Rolf Kalbermatter
Dany Allard [EMAIL PROTECTED] wrote:
 
Is it possible to associate a type of file, like *.txt with Notepad.exe,
with a LV7 app.
I know that LV7 have a new feature that permit to retrieve the command line
used to call the app.
But I don’t think that would be useful for what I want to do. 
I want to Double-Click on my file and my LV7 app start automatically and the
file is read and displayed.
 
You will need the command line too! But first you need to go into the
registry and create the according keys to associate your file ending with
your application. Looking in the registry how things are done for instance
for .vi files should give you enough of an idea.

In the ROOT key you have a key .vi with its value set to some sort of name in
this case LabVIEWInstrument. Under the same root group you get another
group with that LabVIEWInstrument name and some groups and keys in there.
The important one is shell/open/command whose value is the path to your
executable and then %1 as parameter to tell the Windows shell to insert
the file name the user has clicked on at this place.

Then read the command line in your application and do what you need to do
with it.

Rolf Kalbermatter
CIT Engineering Nederland BVtel: +31 (070) 415 9190
Treubstraat 7H   fax: +31 (070) 415 9191
2288 EG Rijswijkhttp://www.citengineering.com
Netherlands mailto:[EMAIL PROTECTED]
 




RE: Dual monitors for labview use

2004-02-13 Thread Scott Serlin
What is the difference in performance I could see between running two
separate pci cards, one pci card and one agp card, or one dual-monitor
card?

-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
Behalf Of Simon Whitaker
Sent: Friday, February 13, 2004 2:59 AM
To: [EMAIL PROTECTED]
Subject: Re: Dual monitors for labview use


Hi Scott,

 Can anyone tell me what graphic cards work well with respect to using
 dual monitors and labview?  How does the graphic card handle moving
the
 display data from one monitor to another. I was considering the ATI
 Radeon 9800XT.  Great for gaming as well. I remember someone talking
 about having the diagram window on one machine and the panel window on
 another.  No more fussing around with the windows to do your work
 faster.

We use dual monitors here, using a variety of different graphics cards.
I have an nVidia GeForce FX5200 (great card, but take care if you buy
one because they do single- and dual-monitor varieties), my boss has a
Matrox card of some description, and one of our machines just uses two
separate single-display cards. That machine runs XP, which has
dual-monitor support built in, so you can use two separate
single-display cards instead of a single dual-display card.

In all cases, LabVIEW works just fine with them. The software we develop
uses a subVI to retain the screen position of VIs and put them back in
the same place next time they're launched (using the VI's Windows.Bounds
property), and that has no problem coping with the dual display setup.
As someone else has noted, XP and Windows 2000 treat dual monitors as
one large desktop, so software doesn't need to be aware of which monitor
it's running on, it just runs somewhere on that large desktop and
neither knows nor cares whether that happens to be on monitor 1 or
monitor 2.

The only noticable difference between the various solutions we use is
the facilities offered by the driver software that comes with the video
card(s). My nVidia card comes with an app called nView that allows you
to do all sorts of fancy things with your dual monitors, such as
enabling window spanning across desktops, choosing which monitor
specific apps open on, setting different wallpapers for each monitor and
so on. The machine that uses two separate cards and relies on the dual
monitor support built into XP lacks these advanced features, although
it's still perfectly usable.

If you have an XP machine with a spare PCI slot, a spare PCI graphics
card and a spare monitor, try plugging the card into the PC and giving
it a go - won't cost you anything, and you'll see it first-hand. Try
before you buy! :-)

 One other off topic question.  Have you guys/gals been buying
 and using LCD monitors?  Anyone running them for a long time?  I keep
 seeing used monitors show up on ebay that have a few/many lcds burnt
out
 or broken in the display.  How hard and expensive are they to repair
if
 this is the case? I can solder as I'm a EE by nature.

I've got an LCD monitor as my 2nd display - a fairly cheap 15
(1024x768)
model that I have pivoted through 90 degrees (another feature supported
by the nVidia software), making it great for working on documents. I've
had it for about 4 months and haven't had any problems so far with dead
pixels. We've got another LCD display that we've had for about a year,
again no problems with dead pixels. Not sure how feasible it is to
repair a dead pixel. Note that even new LCD monitors may ship with some
dead pixels, and most manufacturers specify a tolerance level for dead
pixels below which they won't replace a panel.

All the best,


Simon Whitaker [EMAIL PROTECTED]
Software developer, Tiab Ltd
tel: +44 (0)1295 714046
fax: +44 (0)1295 712334
web: http://www.tiab.co.uk/




Re: tutorial to build CGI application

2004-02-13 Thread Jyoti Fontaine
Hello Philippe,

Here is a link to the CGI tutorial we currently have on our website:
http://zone.ni.com/devzone/conceptd.nsf/webmain/E8A8D4B9497A1E1E86256802007B8E06?opendocument.
It is not step by step, but I think it will help you.

Also, if you need help with building your application into an
executable, please look at the following knowledge base:
http://digital.ni.com/public.nsf/websearch/17427B623FE6083F86256D1300705CFE?OpenDocument

I hope this helps.
Regards,
Jyoti F.
National Instruments

Also, if you need help with



Re: Able to reset the board from MAX but not from my LABVIEW6.1 program

2004-02-13 Thread Dennis Knutson
Jonnie 5 gave you the best advice regarding solving your problem -
post your code. I just wanted to add that when you say Labview
standard program, it doesn't mean an awful lot. There are several
shipping examples of serial communication and they are different
depending on which versioon of LabVIEW that you're using. Real old
ones used the serpdrv and more recent ones use VISA. I don't remember
any of the examples being setup to add a termination character to the
write string. This is one of the most common mistakes people make when
starting to program serial in LabVIEW.



NI reduced prices on some DAQ equipment

2004-02-13 Thread Bookwalter, Dan
I was on NI's website and noticed that they are offering discounts on
several of their popular DAQ devices...

Dan Bookwalter N8DCJ
Hi-Stat a Stoneridge Co.
345 South Mill St.
Lexington, Ohio 44904

(419) 884-1219 Phone
(419) 884-4172 Direct
(419) 884-4195 Fax

[EMAIL PROTECTED]



_ 
This electronic mail transmission contains confidential information 
intended only for the person(s) named. Any use, distribution, copying 
or disclosure by any other person is strictly prohibited. If you 
received this transmission in error, please send an electronic mail 
message to [EMAIL PROTECTED] 





Re: Dual monitors for labview use

2004-02-13 Thread Howard, John
I seem to have more limitations imposed on me by the monitor than the video
card.  According to the G550 specs, it has dual integrated RAMDACs, a 360MHz
Primary and a 230 MHZ Secondary.  It claims up to [EMAIL PROTECTED] on the main
display, but my monitor can't go that fast.  Likewise, the secondary should be
able to go faster than 85Hz @1024x768, but once again the monitor can't.

So maybe your own limitations aren't the video card?  Just something to
consider.

John Howard

 Uwe Frenz 02/13/04 10:37AM 
John,

you wrote:
 I am using a Matrox G550, and it
 works wonderfully! (21,[EMAIL PROTECTED] + 17, [EMAIL PROTECTED])
Thats about or little above the specs, AFAIK.
The highres requires a pixel clock of 160 MHz. The video bandwidth of that
chips is not far beyond 250 MHz, so the signal quality will not be very
impressive. But it works - i use exactly the same resolution on a Nokia
445Xpro.
But the second head used to have a much smaller video bandwidth. The
producers tend to overstate their product capabilities, e.g. allowing
resolutions that do not produce good signal quality - just to have
impressive marketing labels.

-- 
Mit freundlichen Grüßen!
Greetings from Germany!
-- 
Uwe Frenz


~
Dr. Uwe Frenz
Entwicklung
getemed Medizin- und Informationtechnik AG
Oderstr. 59
D-14513 Teltow

Tel.  +49 3328 39 42 0
Fax   +49 3328 39 42 99
[EMAIL PROTECTED] 
www.Getemed.de




Re: Filtering waveform?

2004-02-13 Thread LocalDSP
LabVIEW 6.1 offers some few resampling options with the low level VIs.
(Linear and Spline interpolation). See for example the VI attached to
this reply:

http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=137HOID=5065000500CE580100HTHREAD=88270UCATEGORY_0=_49_%24_6_UCATEGORY_S=0



RE: Serial comms

2004-02-13 Thread Khalid . Ansari

  
Hello Craig,

I would like to recommend PortMon -- a freeware utility to monitor all serial (and 
parallel) port activity with myriad features.  

http://www.sysinternals.com/ntw2k/freeware/portmon.shtml

I've found it to be an invaluable tool for troubleshooting serial comm issues.. in 
fact we joke about it as PortMon -- An Engineer's PokeMon!

Hope you find it equally useful.

Regards,

Khalid Ansari 


- Forwarded by Khalid Ansari on 02/13/2004 11:29 AM -
Craig Graham [EMAIL PROTECTED]
Sent by: [EMAIL PROTECTED]
02/13/2004 06:30 AM
To: [EMAIL PROTECTED]
cc: 
Subject: Serial comms


[clip]
Since I have a laptop and 6036 card lying around, and I have a demo program
for this reader, I think I'm going to end up doing a simple serial analyser
and seeing what the differences are between my code and the demo. But that's
not going to happen till next week and it will waste quite a bit of time so
I thought I'd see if anyone had suggestions.
[clip]

  
  




RE: Dual monitors for labview use

2004-02-13 Thread Scott Hannahs
At 10:34 -0600 02/13/2004, Scott Serlin wrote:
What is the difference in performance I could see between running two
separate pci cards, one pci card and one agp card, or one dual-monitor
card?

Depends on what else you use the PCI bus for.  It is slower performance for raw bit 
pushing than the AGP so if you have a smaller monitor use that one on the PCI bus.  If 
you are also trying to acquire 60 MS/s of data, then I wouldn't recommend the PCI bus. 
 It is a limited resource like everything else.




Re: Main VI stops when using control references.

2004-02-13 Thread Joe Guo
Assuming you use the VI server to load the subvi dynamically, when
loading the vi, the Wait to finish must be set to false, so your
main loop does not wait for the subvi to finish.  When the main loop
is waiting for the subvi to finish, that case (loop) is halted.

Joe



Using Tabs with Subvi's

2004-02-13 Thread Heathman
I wish to have each page of a tab control be used with a subvi. One
for graphs, one for data entry,etc and wish to have the tab page
number start and stop the selected subvi. Up until now I  have been
using 'run on open' and a seperate stop button for each subvi. I tried
creating dummy tab pages and local variables and it works but is
clunky. The problem is that I can't read a tab page number selected
from a main program while I'm in a subvi. Thanks, Jim W



Re: Labview PDA using RS232

2004-02-13 Thread techserve
I was looking for a solution to this question!



Re: Why does converting a SGL to SGL get rid of coersion dots?

2004-02-13 Thread Jeremy C
The algorithm for determining resulting data types gets much more
complicated when using uninitialized shift registers.  The algorithm
used by the binary arithmetic functions is slightly different than the
algorithm used by the compound arithmetic function.  It was determined
that the algorithm used by the compound arithmetic function is more
robust than that used by the binary arithmetic functions.
Modification to the algorithm used by the binary arithmetic functions
is being considered for future versions of LabVIEW.



Re: Problem with DAQMX digital output

2004-02-13 Thread rgharrison
Hi V.GRAF,

The DAQ 6025E is a special card in terms of digital I/O.  This is
because it has two different semiconductor chips handling the digital.
It has the DAQ-STC chip handling the first port of 8 bits (Port 0).
It then has the 8255 chip handling the 3 other ports.

In general, you should be able to use any of the DAQmx programs for
digital with this card.  However, you have to make sure that you are
configuring the entire port at the beginning of your program. This is
because if you reconfigure any of the 3 ports on the 8255 PPI chip, it
will reset the port.  So my suggestion is to not use line
configuring but instead use port configuring.

Here are a couple Knowledge Bases that discuss the issue.

The NI PCI-6025E Digital Line Virtual Channels Return Error -10007 in
Measurement  Automation Explorer Test Panel
http://digital.ni.com/public.nsf/websearch/77DBEDFCCA418A8DC1256C16007FA3B7?OpenDocument
Why Do I Lose the Data on Ports 2, 3, and 4 of a PCI-6025E When
Reconfiguring Port 0 in Visual Basic?
http://digital.ni.com/public.nsf/websearch/33BE87B9752B7D7186256B1E0050A1B9?OpenDocument

Hope that helps.  Have a good day.

Ron
Applications Engineering
National Instruments



Re: Sequence tunnel error

2004-02-13 Thread JoeDoe
Ok guys here where I stand. I tried to pull the significant parts of
the code out and put them in a new VI soo that i could post it to the
board, but When i did so the new VI worked flawlessly. So i toyed
around more removin parts of the sequence and different frames. I have
narrowed it down to one specific area. Once the data was inside the
sequence it goes to a subvi that is used for updating. This subvi
output is the data is then fed to the hardware and fed to that tunnel
thats giving me the trouble. Ok when the data leaves that subvi it
goes to the output tunnel but it also goes to a sequence local to be
used in a latter frame. This is the part thats causeing the change of
data. If i remove that sequence local and instead use a local
variable( i know another no no..) it works fine. Its as if the
sequence local somehow interacts with the seqence tunnel. Im not sure
if these data values are stored in the same memory spot or what but
having that sequence local and the sequence tunnel have the same data
value seems to screw everything up. For now im just going to use the
local variable and be happy i have a work around. If i can figure out
how to duplicate the problem ill post the code. I would post the
entire VI but its got alot of hardware funcionality that im not sure
what it would do to a system with out the hardware. Thanks for
offering suggestions, they did lead me to a fix.



Re: subPanel: The coordinate system seems all scr... up? AND How...

2004-02-13 Thread David Duffey
The top-left corner of the subPanel uses the same coords as the
top-left corner of your saved sub VI.  Also the subPanel document
size will be the same size as your running VI.  I.E. the sub VI's
should have the same placement / appearance as regular VIs, or
exsiting in a subPanel.

If you want to make sure they match up, then save your sub VIs with
the origin at the top-left (0,0).

One way to do this is to place a property node on your block diagram,
right click the header to change the property class to VI and then
select origin property, right click to change to write.  Right click
again to create a cluster (0,0).

When you run your VI, the top-left will align perfectly with the two
thick lines.  Save your VI so that the VI will open at this coord
(even in the subPanel).

-David



NI SEC Invoke Loging Dialog

2004-02-13 Thread Ezequiel
I=B4m using a NI SEC Invoke Loging Dialog.vi in my application but when
I compile it in a stand alone .exe file, the NI SEC Invoke Loging
Dialog doesn=B4t work.



Re: Some Problems with LabVIEW and Matlab DLLs

2004-02-13 Thread ColinAE
Hello spj.

I have taken a look at your issue.  It sounds like what is happening
could be expected behavior.  When you re-build your script into a DLL,
as is mentioned in your post, it will re-register the COM object to
Windows.  Therefore, even though the names and locations are the same,
the DLL (really like an .ocx file in this case) has been re-registered
and thus will need to be re-linked to the LabVIEW Automation Open
function each time.

I hope this helps!
Colin C.
Applications Engineer



Re: How to compile wavread on MatLab Script?

2004-02-13 Thread Zvezdana S.
Janet,
Remove the quotes around blah.
Just put in your MATLAB Script Node
f=wavread(blah)
With the quotes, MATLAB was trying to read the string 'blah', and not
the blah input, which is 'annano.wav'

Zvezdana S.
National Instruments



Re: 34970A VI?

2004-02-13 Thread Jonnie 5
If you're talking about an HP/Agilant, there are vi's included with
the driver:
http://zone.ni.com/idnet97.nsf/9b2b33e1993d877786256436006ec498/e4c553c139f78062862568ab005fba3c?OpenDocument

See also:
http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=50650008001786UCATEGORY_0=_49_%24_6_UCATEGORY_S=0USEARCHCONTEXT_QUESTION_0=34970AUSEARCHCONTEXT_QUESTION_S=0

Or just search the discussion forums for 34970A...



Re: How do you reset the encoder count with LV?

2004-02-13 Thread Bates
STK,

Yes Rest Encoder Count is for resetting a unmapped encoder.  For an
encoder that is mapped use Reset Position  This will change the
value of the encoder that is mapped to an axis.

A. Talley
Application Engineer
National Instruments



Re: Unable to lock memory chunk, unable to allocate memory

2004-02-13 Thread Sean C.
Hello Matute,

Thank you for contacting National Instruments.

Try looking at the discussion thread linked below.  It concerns a
problem similar to yours.

http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RNAME=ViewQuestionHOID=5065000800BEA7ECategory=LabVIEW.LabVIEW+General

Let me know if this does not help.

Sean C.
Applications Engineer
National Instruments



Re: Write hex code to pic with serial port, any example?

2004-02-13 Thread Sean C.
Hello Jose,

Thank you for contacting National Instruments.

The following link leads to a discussion thread about hex to ASCII
serial communication.

http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RNAME=ViewQuestionHOID=5065000800F544ECategory=Instrument+Connectivity.Serial+Interfaces

Please reply to this post if you have some more specific questions.

I hope this helps,

Sean C.
Applications Engineer
National Instruments



Re: Sequence tunnel error

2004-02-13 Thread Dennis Knutson
One of the problems with the sequence structure is that data doesn't
leave any of the frames until the last frame is done. I suspect that
that is what you are seeing if the data is modifed in a later frame.



Re: how does the interface to ADO drivers work?

2004-02-13 Thread mikeporter
I'm using the ActiveX interface.

Mike...



low level device drivers

2004-02-13 Thread Howard, John
OK, I think I am in over my head.  :)

I am considering using a PCI card (Sabtech NTDS interface card for those
interested), and it doesn't seem to have a nice simple DLL like I am use to. 
Instead it has a kernel mode Windows 2000 device driver, and I am expected to
use ...normal Windows I/O system calls... to talk to it.  The driver is
written to be part of the Serial Port driver class, BUT it can not be treated
simply as a serial port since ...there are several restrictions on the I/O
calls..., and ...it has its own set of DeviceIoControl functions.

I know so little about Windows device drivers that I don't really even
understand the paragraph I just wrote.  Does this mean I can't talk to the
card directly from LabVIEW?  If I can, is it difficult for someone who uses
DLL's fairly often?  Can I use the Win32 API to talk to it?  Does anyone have
any suggestions on how to approach this?

There is another card I am considering from a different company that does have
a DLL (with a rather obscure command set), which I could use instead of this
one.  Or maybe I could get one of our .NET programers to write me something
simple.  Suggestions anyone?

Thanks for any nuggets of wisdom!

John Howard




Re: How can i display the signal out of a FFT block to a graph?

2004-02-13 Thread Jyoti Fontaine
Hello St. Augustine,

You should not need to do any conversion from the output of the FFT
VIs to the waveform graph.  Make sure, however, that you are wiring
into a Waveform Graph.vi and not the Waveform Chart.vi.

Regards,
Jyoti F.
National Instruments



Re: how can I create a simulated hardware configuration in MAX Measurement Automation Explorer

2004-02-13 Thread Jonnie 5
There really isn't anyway to fake having a DAQ board, and w/o a DAQ
board, you're going to get errors when trying to communicate to it.

You can either:
1) Remove the hardware commands and use simulated dsta.
http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800A162UCATEGORY_0=_49_%24_6_UCATEGORY_S=0

2) Use an alternative DAQ such as a serial/parallel port or even a
sound card
http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800C7B2UCATEGORY_0=_49_%24_6_UCATEGORY_S=0USEARCHCONTEXT_QUESTION_0=simulatedUSEARCHCONTEXT_QUESTION_S=0

3) Connect to the PC w/ the DAQ remotely to test the program (Remote
Data Acquisition or RDA).
http://exchange.ni.com/servlet/ProcessRequest?RHIVEID=101RPAGEID=135HOID=5065000800E76EUCATEGORY_0=_30_%24_12_UCATEGORY_S=0USEARCHCONTEXT_TIER_0=0USEARCHCONTEXT_TIER_S=0USEARCHCONTEXT_QUESTION_0=simulatedUSEARCHCONTEXT_QUESTION_S=0



Re: Main VI stops when using control references.

2004-02-13 Thread Dr. Imad
the Engine Momentum Subvi should finish execution with each iteration
of the loop in the DAQ.vi
Therefore , connect a constant Boolean set to True to the condiotional
terminal of the while Loop of the Engine Momentum Subvi (instead of
the stop value from the main vi)



Re: Main VI stops when using control references.

2004-02-13 Thread mkdcs
I tried that and it works but the zero outside the loop which resets
the shift register to zero each time the loop starts must be removed
so the engine accelerates.  I just need to figure out how to reset the
speed to zero after I've stopped and restarted the main vi without
have to close it and reopen.

Thanks,
Mike



Re: Read Trajectory Data Function and 7344 card

2004-02-13 Thread Bates
Chris,

The difference is the hardware 3ms is a 7350 and 7344 is the 10 ms.
The diference is which card you by.

A Talley
Applications Engineering
National Instruments



RE: Dual monitors for labview use

2004-02-13 Thread Scott Serlin
Can this feature, two independent desktops be implemented in win2k or
winXP?  Also, regarding the potential invisible area on the screen,
can't you just go to a small resolution on the larger screen so that the
two screen match?  I have to commend this list as I would never have
thought about a dual-monitor setup if one poster (who I have
forgot-sorry) had not mentioned it. I was on a website recently and they
had a rack to set up multiple LCD screens on one monitor rack.  I wonder
what the video card would look like if trying to drive 3 or more
monitors.  What's out there just for grins?  How expensive?  I could
have my labview diagram on one screen, panel on the other, dvd video
playing on a third, etc.  I just need to install more eyes into my head
to view all of this.  8-)

On Win9x however there will be just two independend desktops

-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
Behalf Of Uwe Frenz
Sent: Friday, February 13, 2004 1:50 AM
To: Scott Serlin
Cc: LV-Info, list
Subject: Re: Dual monitors for labview use


Scott,

you asked on Thu, 12 Feb 2004 13:26:27 -0600:
 Can anyone tell me what graphic cards work well with respect to using
 dual monitors and labview?  How does the graphic card handle moving
the
 display data from one monitor to another. I was considering the ATI
 Radeon 9800XT.  Great for gaming as well. I remember someone talking
 about having the diagram window on one machine and the panel window on
 another.  No more fussing around with the windows to do your work
 faster.  
Well, here at getemed most of the devellopers have now a dual monitor
system. Most of us have matrox G550s, but almost all dual monitor cards
should work well too.

I had a project where the customer even wanted three large monitors. So
I
added two graphics cards, an ATI and a Matrox dualhead into the machine.
All worked well, allthough (because of the 3 monitors) I had to use
Win98
at that time.

Here we come to a big caution:
Win NT  2k (no knowledge on XP yet) handle multiple monitors as two
windows to a large and unique desktop, whereas Win9x used to have
multiple
independend desktops that just had to share one side with each other.
This
might be important, because almost all dualhead cards I know of have an
excellent head as primary head, supporting  video bandwidth of up to
250
MHz (and resolutions of up to 1600*1200). The secondary head is propably
less capabal, supportimg only smaller resolutions. This may result in
some
area on the desktop that is not displayed on any of the screens in W2k.
Example: One screen with 1600*1200 left and another with 1280*1024 right
aligned at bottom. The desktop will be (1600+1280)*1200 and there will
be
an area of (1600, 1199) to (2879, 1024) that will be invisibel.
On Win9x however there will be just two independend desktops.

 Also, how about those kvm monitor switches?  Anyone used those
 in conjunction with the dual monitor graphic cards?  Is going with an
 extra PCI board as well as the ATI Radeon 9800XT AGP board the better
 route than running both monitors off of the same card?  Thoughts?
 Comments?  
Monitor switches have to support the high video bandwidth in order not
to
damage signal quality. Try it out for the equipment and resolution you
are
trying to use. 

 One other off topic question.  Have you guys/gals been buying
 and using LCD monitors?  Anyone running them for a long time?  I keep
 seeing used monitors show up on ebay that have a few/many lcds burnt
out
 or broken in the display.  How hard and expensive are they to repair
if
 this is the case? I can solder as I'm a EE by nature.
AFAIK this is an internal damage that can't be repaired and never ever
with
a 'burning iron'.
Nothing to say about your 'natural' solder art crafting.
(Hope this is got as that joke it is intended to be. Joking in a foreigh
language is not so easy.)

The LCD producers have even defined a level of defective pixels that are
considered as OK according to the production standards. Otherwise they
would get too few panels being OK out of their production lines and
those
would be too expensive for mass market.

Greetings from Germany!
-- 
Uwe Frenz


~
Dr. Uwe Frenz
Entwicklung
getemed Medizin- und Informationtechnik AG
Oderstr. 59
D-14513 Teltow

Tel.  +49 3328 39 42 0
Fax   +49 3328 39 42 99
[EMAIL PROTECTED]
WWW.Getemed.de




Re: subPanel: The coordinate system seems all scr... up? AND How...

2004-02-13 Thread Skinny
Thanks!

- Martin



Serial Port

2004-02-13 Thread Subramanian, Shankar, K
Hi

I have done data acquisition using NI DAQ card for one of my projects. Right
now, have a small project in which data needs to be acquired from a serial
port. The system has a temperature controller unit which sends data to the
serial port and needs to be collected using LabView. 

How different is it to configure and acquire data using serial port rather
than NIDAQ card. Could anyone help me out in this? Also, this is going to be
done using LabView 7 - Evaluation Version. I guess there are no restrictions
while we use evaluation version. I hope all the features of a full
development system could be used. 

Thanks for the help, 

Shankar 




scaling graph to frequency?

2004-02-13 Thread St Augustine
i have succesfully read in data from a digital input and managed to
store in a text file. This data appears to be most likely an ASCII
format. Are there ASCII to number converters? I would also like to now
use this data and perform analysis on them.For example using the FFT
etc. I have managed to read the data from the same file and perform
analysis on it.However the FFT must represents a signal in the
frequency domain and the scale remains in time. Im not sure how to
scale my X axis to represent this? Can you please help!



RE: Dual monitors for labview use

2004-02-13 Thread kwhite




On Fri, 13 Feb 2004, Scott Serlin wrote:

 What is the difference in performance I could see between running two
 separate pci cards, one pci card and one agp card, or one dual-monitor
 card?




most likely, zero difference if you by a recent agp dual head card.
possibly fewer driver issues with a single card as well.

-k




Can we acquire the Analog input and Digital Input at

2004-02-13 Thread chenchen
Can we acquire the Analog input and Digital Input at
the same time? My sensor data include Analog and
digital signals.



RS232 Displays for LVRT

2004-02-13 Thread Sergey Liberman
Hi all,

Does anyone know of any display (preferably touchscreen, but not
necessarily) that can be controlled over RS232 interface (rather than a
video board or chipset)? I need to connect a local display to a system
running LabVIEW RT, and it looks like RS232 is the only option that I have.
I don't need a VGA resolution - smaller resolution, black-and-white LCD will
be fine. I know there was at least one company that had products like that -
it had a booth at NIWeek a couple years ago - but I can't recall its name.
TIA,

   Sergey Liberman   [EMAIL PROTECTED]
   Solidus Integration   http://solidusintegration.com