Hello,
I am using JMF (Java Media Framework) to video stream, and then i am taking a
snap shot of that video stream at every 5 seconds. there is a class called
FrameGrabber in JMF that grabs a frame and return that frame as java.awt.Image.
Then i took that image and deinterlace it.
The only thing change in the above method is the dataBuffer, the byte[] data
changes in every loop from the images from the video. The dimension is the
same. Is there a way to reset the dataBuffer without the recalling new
DataBufferByte(dataBuffer, width * height)? i saw a data buffer that
First, i am extracting images from the video stream by taking a snap shot of
each images from the video, and then I extracted the pixels data from those
images using pixelGrabbers. This return a byte[] array of pixels information of
the image. then i create BufferedImage using the method i
How are you taking the snap shot? How do you receive that snapshot into
Java? By loading it as an image, or are you using the built-in robot
facilities to do the snapshot?
If you are using robot, then the image will already be a BufferedImage
so you don't need to convert it. If you are
Your dataBuffer array will by far be the biggest problem, the rest of the
objects don't take much time - they're just wrappers allowing access to the
data array.
You might want to cycle between a couple of data arrays or something instead of
creating a new one every time.
Dmitri
[Message sent
the rest of the objects don't take much time
And by time I mean space. They're the same thing anyway!
Dmitri
[EMAIL PROTECTED] wrote:
Your dataBuffer array will by far be the biggest problem, the rest of the
objects don't take much time - they're just wrappers allowing access to the
What do u mean by cycling two data array? and how would i get started on that?
is there a some sort of method on the dataBuffer that i can reset my byte[]
array?
thank you
Francis
[Message sent by forum member 'cohodetector' (cohodetector)]
I don't have enough information on what exactly you're doing to advise you on
the implementation, but I thought that if you have something that generates a
new array of pixels on every frame (may be from a different thread or
something), may be you could have two arrays - one you currently
Hello,
I have a method that creates a BufferedImage from a byte[]. It is a gray-scale
image, and each pixels in the image is represent by a single byte range from 0
to 255. here is the method:
public BufferedImage produceRenderedImage(byte[] dataBuffer, int width, int
height)
{