Hi Jack,

Thanks for your suggestions! It’s taken me a little while to work through it 
all.

I did manage to get a 2-pol, 32k channel design to compile eventually. I think 
it was a combination of a) sufficient latency in the Vacc and PFB blocks (as 
you suggest) and b) splitting the two pols into two separate chains, rather 
than using one block for the PFB and FFT in biplex mode. It was found by just 
iterating through latency combinations until the timing started to converge, 
but there were still several hundred errors until I split up the design into 
two chains. Now, I don’t really understand why using two separate PFB/FFT 
blocks should make a difference (surely sysgen sorts such detail?), but the 
latency settings were:

PFB: Add Latency 1; Mult latency 2; BRAM latency 3; Fanout latency 1; Convert 
latency 1 (4 taps).
Vacc: Add latency 2; BRAM latency 6 (overkill?); Mux latency 0,

where DSP48 has been used wherever possible in the design.

Now I need to see if it still works if I use more FIR taps ;-)

Best wishes
Michael

From: Jack Hickish [mailto:jackhick...@gmail.com]
Sent: 30 August 2015 21:57
To: Michael D'Cruze
Cc: casper@lists.berkeley.edu
Subject: Re: [casper] Mystery timing errors

Hi Michael,

As one of the Xilinx timing closure documents helpfully articulates, it's very 
difficult to give specific recipes for solving timing problems, other than 
reading the timing report, looking at things in PlanAhead/FPGA Editor, 
iterating compiles and developing some intuition.

That said, there are a few things you could try, of varying levels of 
complexity --
Since you have quite a lot of channels in your design, I suspect your vacc 
problems are related to the size of the bram buffer (especially if your samples 
are wide). Keep in mind that a bram in a Virtex6 is a 36 bit wide x 1k deep 
memory block, so building a vacc to accommodate many thousands of channels 
means banking together lots of brams -- this is liable to cause fanout issues 
on your address/control signals. The same is probably true of the large delays 
in the PFB blocks, though I'm not sure how these are implemented in the various 
block versions which exist in mlib_devel.
Setting brams to optimize for speed and giving them at least 3 cycles latency 
may help. You could also try manually controlling bram control signals -- I 
believe the casper_library_bus/bus_single_port_ram block will do some of this 
for you if you set to optimize for speed and include some fanout latency. If 
you replace the bram delay in the vacc with one of these (and associated 
address counter logic to make the block act as a delay) then perhaps this will 
help. Be careful to get your latencies right so the block still functions 
properly.
You might also find that implementing the adder in a DSP slice (if you haven't 
already) might help.

A more involved option is to go about manually constraining the placement of 
components. You might find setting up pblocks in PlanAhead to place major parts 
of your design might free up the compiler to do a little better with the 
remainder of the logic. Adding pblocks for the pfb, and the various FFT stages 
can be very helpful in this regard. Or perhaps just constraining the vacc 
that's causing you problems might be enough.
Once you have some constraints from planahead which work, you can auto-include 
them in your simulink compiles using various methods, my favourite being the 
'UCF' yellow block, which is in the current casper-astro mlib_devel.

In general, I suspect that you would be well served by trying to reduce your 
bram use in your model - the resource utilisation report can probably either 
confirm or contest that this is the case. You could do this by trying to use 
alternatives to brams where blocks allow it (for example the FFT will allow 
coefficients to be generated using DSPs in some cases), or by reducing the 
number of FIR taps, or implementing something using QDR rather than bram if 
appropriate.

The unfortunate truth is that some, none or all of these suggestions may help, 
or might make things worse. I would say that in my experience over-judicious 
use of pipeline/register stages throughout a design can often do more harm than 
good, and a sensibly placed PFB can result in incredible timing improvements.

Hope that helps, please email back with any more info and/or updates -- I 
suspect if you solve this problem documenting your method may prove very useful 
for others encountering similar problems.

Cheers, and good luck!
Jack

On 30 August 2015 at 10:42, Michael D'Cruze 
<michael.dcr...@postgrad.manchester.ac.uk<mailto:michael.dcr...@postgrad.manchester.ac.uk>>
 wrote:
>
> Hi everyone,
>
>
>
> I’m having quite a lot of problems getting a particular design to compile, 
> XPS consistently reporting timing errors. The design is a wideband 
> spectrometer, somewhat similar to the tutorial 3 design. I’m running an iADC 
> at 1024 MHz, and the FPGA at 256 MHz. It is a two-polarisation design, with a 
> 64k-point PFB and FFT (for 32k channels) in each polarisation. I am running 
> the PFB and FFT blocks in “two-polarisation” mode, so there are just one of 
> each block. I am using the Vacc block from the xBlocks repository.
>
>
>
> I should add at this point that the same design with 16k channels compiles 
> successfully at this speed, and that the 32k channel design compiles at 
> slower speeds (e.g. 200 MHz FPGA).
>
>
>
> The timing reports suggest negative slack in the PFB and Vacc blocks. I’ve 
> tried adding various combinations of latency in each, however no permutations 
> have resulted in substantial improvements. The overwhelming majority of the 
> errors are reported by the Vacc block. I have also tried replacing the 
> xBlocks Vacc block with the wide_bram_vacc block in 64-bit mode, however this 
> results in even more timing errors (and without the option to adjust 
> latencies).
>
>
>
> I’ve tried various combinations of latencies suggested previously on the mail 
> archive, but again nothing has really given any improvement.
>
>
>
> Am I simply pushing such a large design too hard? Is the only option to slow 
> it down or is there some strategic method I can adopt to get closer to timing 
> closure? Suggestions much appreciated!
>
>
>
> Thanks
>
> Michael

Reply via email to