[Pafgbt] PAF beamformer size and cost

Rick Fisher rfisher at nrao.edu
Thu Feb 4 14:44:38 EST 2010


What are the Arecibo 7-beam pulsar survey parameters?

Rick

On Thu, 4 Feb 2010, Scott Ransom wrote:

> For L-band with 500MHz of BW, you'd probably want 2K channels (maybe 1K)
> dumping at between 50-100us.
>
> Scott
>
>
> On Thursday 04 February 2010 02:35:57 pm Rick Fisher wrote:
>> John,
>>
>> How many frequency channels are there and what's the time resolution of
>> guppi in search mode?
>>
>> Rick
>>
>> On Thu, 4 Feb 2010, John Ford wrote:
>>>> Brian,
>>>>
>>>> Thanks very much for the beamformer outline.  How does Jason handle
>>>> the frequency dependence of the beamformer weights?  Is there an FIR
>>>> is each signal path or maybe an FFT - iFFT operation?
>>>>
>>>> What to do with the high bandwidth beam outputs at high time
>>>> resolution is a standard pulsar problem.  Since the beam outputs are
>>>> essentially the same as you'd have with a conventional horn feed,
>>>> all of the current tricks of the trade will apply.  As far as I
>>>> know, all pulsar surveys are done by generating spectra with a
>>>> frequency resolution consistent with the highest expected pulse
>>>> dispersion, squaring the signals, integrating for the selected
>>>> resolution time, and streaming this to disk.  Pulse searches are
>>>> then done off-line in computer clusters or supercomputers.  (Our
>>>> pulsar experts can chime in here.)  We'll need to match our output
>>>> data rates with available storage and post-processing capabilities -
>>>> a time-dependent target.  Maybe someone could give us some currently
>>>> feasible numbers and time derivatives.
>>>
>>> I'm not a pulsar expert, but I know a bit about what the 1 beam GBT
>>> currently can do, and what is available in the computer world.
>>>
>>> I don't think it's feasible to keep 40 beams of pulsar search data.
>>> We're having a time of it keeping up with managing the data from our
>>> 1 beam guppi.  We capture ~50-100 MB/sec with guppi in search mode,
>>> so 40 times that is completely insane. (20-40 Gigabytes per second!)
>>>
>>> It's not so much the capture rate (which, while difficult, is
>>> possible today with large disk arrays) as the fact that pulsar
>>> searches use long integrations, so the volume of data to be stored
>>> until it is processed is huge.  This example would be >500 Terabytes
>>> in an 8 hour observation. Yikes!  That's possibly more pulsar search
>>> data than has been collected, ever.
>>>
>>> I have no idea how long it takes to process the data, but the only
>>> really feasible thing to do for this kind of data rate is to process
>>> it in (near) real time and throw it away when you are done.  That
>>> means a *really* large supercomputer or special-purpose pulsar search
>>> engine.
>>>
>>> Pulsar timing is another kettle of fish altogether.  There, the
>>> bottleneck is processing power. (assuming coherent dedispersion)  The
>>> disk I/O is almost negligible.  The I/O requirements for getting the
>>> samples off the beamformer and onto the processing machine are also
>>> formidable.
>>>
>>> As usual, the pulsar requirements far outstrip all other observing
>>> requirements as far as data rates.  We can build spectrometers that
>>> integrate for milliseconds or seconds to cut data rates to manageable
>>> numbers.  So it seems reasonable to build a machine that could handle
>>> the 40 beams for "normal" observing, and reduce the number of beams
>>> used for pulsar search output.
>>>
>>> For transients and pulsars you need the fast-sampled time-domain data
>>> to work with, and can't really integrate it down much before using
>>> it.
>>>
>>> I think the biggest problem is really storage of the data, and not
>>> necessarily the processing in real time.
>>>
>>> John
>>>
>>>> Before completely buying into the voltage sum real-time beamformer
>>>> we should keep in mind that a lot of single dish applications don't
>>>> need voltage outputs as long as the time and frequency resolution
>>>> parameters are satisfied.  If there are big computational savings in
>>>> a post-correlation beamformer, and we satisfy ourselves that there's
>>>> not a hidden gotcha in this approach, we should keep it on the
>>>> table.  My guess is that any computational advantages evaporate or
>>>> even reverse when the required time resolution approaches the
>>>> inverse required frequency resolution.
>>>>
>>>> Rick
>>>>
>>>> On Thu, 4 Feb 2010, Brian Jeffs wrote:
>>>>> Rick,
>>>>>
>>>>> See below:
>>>>>> Is your assumed beamformer architecture voltage sums or
>>>>>> post-correlation?
>>>>>> In other words, are the beams formed by summing complex weighted
>>>>>> voltages
>>>>>> from the array elements or by combining cross products of all of
>>>>>> the elements?  John's reference at
>>>>>> http://arxiv.org/abs/0912.0380v1 shows a voltage-sum beamformer.
>>>>>> The post-correlaion bamformer may use fewer processing resources,
>>>>>> but it precludes further coherent signal processing
>>>>>> of each beam.
>>>>>
>>>>> Our plans are based on a correlator/beamformer developed by Jason
>>>>> Manley for
>>>>> the ATA and some other users (the pocket packetized correlator).
>>>>> He recently
>>>>> added simultaneous beamforming to the existing correlator gateware,
>>>>> so they
>>>>> run concurrently.  In our application the only time this is
>>>>> required is during interference mitigation.  Normally we correlate
>>>>> during calibration and
>>>>> beamform otherwise.
>>>>>
>>>>> His design is a voltage sum real-time beamformer.  At this point he
>>>>> does not
>>>>> compute as many simultaneous beams as we need to, so I think we
>>>>> will have to
>>>>> exploit the computational trade-off to do either beamforming or
>>>>> correlation
>>>>> but not both, or it will not fit in the FPGA.  Post-correlation
>>>>> beamforming
>>>>> is really quite trivial, and has a low computational burden, so
>>>>> that could be
>>>>> added to the correlator and run simultaneously.  I believe that
>>>>> when we need
>>>>> simultaneous voltage sum beamforming and correlations (as when
>>>>> doing interference mitigation) we will have to reduce the effective
>>>>> bandwidth. We
>>>>> really cannot take Jasons' existing code and plug it right in for
>>>>> our application, but it will serve as a very good template.  That
>>>>> is why we have
>>>>> Jonathan out at UC Berkeley for 6 months, so he can learn the ropes
>>>>> and then
>>>>> work on our correlator/beamformer.
>>>>>
>>>>>> Very roughly, the science requirements for a beamformer fall into
>>>>>> two camps, which may be operational definitions of first science
>>>>>> and cadallac/dream machine: 1. spectral line surveys with
>>>>>> bandwidths in the 3-100 MHz range and very modest time resolution
>>>>>> and 2. pulsar and fast transient source surveys with bandwidths on
>>>>>> the order of 500+ MHz and <=50
>>>>>> microsecond time resolution.  The 2001 science case says pulsar
>>>>>> work requires bandwidths of 200+ MHz, but the bar has gone higher
>>>>>> in the meantime.  One can always think of something to do with a
>>>>>> wide bandwidth,
>>>>>> low time resolution beamformer, but it would be a stretch.  The
>>>>>> GBT sensitivity isn't high enough to see HI at redshifts below,
>>>>>> say, 1350 MHz
>>>>>> in a very wide-area survey.  Hence, building a beamformer with
>>>>>> wide bandwith but low time resolution may not be the optimum use
>>>>>> of resources.
>>>>>> Also, the 2001 science cases assumes 7 formed beams, but the
>>>>>> minimum now
>>>>>> would be, maybe, 19 and growing as the competition heats up.
>>>>>
>>>>> We are operating under the assumption of at least 19, and probably
>>>>> more than
>>>>> 40 formed beams.  If we only use the correlator for calibration,
>>>>> then we should be able to achieve both relatively wide bandwidth
>>>>> (250 MHz) and high
>>>>> time resolution (we will get a beamformer output per time sample,
>>>>> not just
>>>>> per STI interval).  Dan and Jason feed that based on their
>>>>> experience with
>>>>> existing codes this is achievable on the 40 ROACH system we
>>>>> sketched out, but
>>>>> we will have to wait and see.  If we run into bottlenecks we will
>>>>> have to
>>>>> reduce either bandwidth or the number of formed beams.
>>>>>
>>>>> One issue I am not clear on yet is what we do with the data streams
>>>>> for 40+
>>>>> voltage sum beams over 500+ frequency channels.  How do we get it
>>>>> off the
>>>>> CASPER array, and what will be done with it?  For 8 bit complex
>>>>> samples at
>>>>> the beamformer outputs you would need the equivalent of fourty 10
>>>>> Gbit ethernet links to some other big processor, such as a
>>>>> transient detector. If
>>>>> this is unreasonable then either the number of bits per sample,
>>>>> bandwidth, or
>>>>> number of beams will need to be reduced.  Alternatively, it is not
>>>>> hard to
>>>>> add a spectrometer to the beamformer outputs inside the
>>>>> correlator/beamformer, and this provides a huge data rate
>>>>> reduction. But how
>>>>> do we handle data for transient observations where fine time
>>>>> resolution is
>>>>> critical?
>>>>>
>>>>> Brian
>>>>>
>>>>>> Counter-thoughts?
>>>>>>
>>>>>> Rick
>>>>>>
>>>>>> On Wed, 3 Feb 2010, Brian Jeffs wrote:
>>>>>>> Rick,
>>>>>>>
>>>>>>> We have a rough architecture and cost estimate for a 40 channel
>>>>>>> correlator/beamformer capable of 40 channels (19 dual pol
>>>>>>> antennas
>>>>>>
>>>>>> plus
>>>>>>
>>>>>>> reference or RFI auxiliary) over 250 MHz BW.  We worked this out
>>>>>>> with CASOER
>>>>>>> head Dan Werthimer and his crack correlator/beamformer developer
>>>>>>
>>>>>> Jason
>>>>>>
>>>>>>> Manley.  It will require 20 ROACH boards, 20 iADC boards, 1
>>>>>>> 20-port
>>>>>>
>>>>>> 10
>>>>>>
>>>>>>> Gbit
>>>>>>> ethernet switch, and some lesser associated parts.
>>>>>>>
>>>>>>> Our recent ROACH order was $2750 each, iADC: $1300 each,
>>>>>>> enclosures:
>>>>>>
>>>>>> $750
>>>>>>
>>>>>>> each, XiLinx chip: free or $3000,  ethernet switch: $12000.
>>>>>>>
>>>>>>> You can use your existing data acquisition array of PCs as the
>>>>>>> stream-to-disk
>>>>>>> farm, but will need to buy 10 Gbit cards and hardware RAID
>>>>>>
>>>>>> controllers.
>>>>>>
>>>>>>> The total (which will be a bit low) assuming no free XiLinx parts
>>>>>>> and
>>>>>>
>>>>>> not
>>>>>>
>>>>>>> including  is:  $168,000.
>>>>>>>
>>>>>>> Of course this does not include development manpower costs.
>>>>>>>
>>>>>>> Brian
>>>>>>>
>>>>>>> On Feb 3, 2010, at 3:05 PM, Rick Fisher wrote:
>>>>>>>> This is an incomplete question, but maybe we can beat it into
>>>>>>
>>>>>> something
>>>>>>
>>>>>>>> answerable:  Do we know enough about existing applications on
>>>>>>
>>>>>> CASPER
>>>>>>
>>>>>>>> hardware to make a conservative estimate of what it would cost
>>>>>>>> to
>>>>>>
>>>>>> build
>>>>>>
>>>>>>>> a
>>>>>>>> PAF beamformer with a given set of specs?  I'm looking for at
>>>>>>>> least
>>>>>>
>>>>>> two
>>>>>>
>>>>>>>> estimates.  What is a realistic set of specs for the first
>>>>>>>> science
>>>>>>
>>>>>> PAF
>>>>>>
>>>>>>>> beamformer, and what would the dream machine that would make a
>>>>>>>> big scientific impact cost?  You're welcome to define the specs
>>>>>>>> that go with
>>>>>>>> either of these two questions or I'll start defining them by
>>>>>>
>>>>>> thinking
>>>>>>
>>>>>>>> "out
>>>>>>>> loud".  The first science beamformer will guide the initial
>>>>>>>> system design,
>>>>>>>> and the dream machine will help get a handle on longer range
>>>>>>>> expectations.
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>> Rick
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Pafgbt mailing list
>>>>>>>> Pafgbt at listmgr.cv.nrao.edu
>>>>>>>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>>>>>>
>>>>>> _______________________________________________
>>>>>> Pafgbt mailing list
>>>>>> Pafgbt at listmgr.cv.nrao.edu
>>>>>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>>>>
>>>> _______________________________________________
>>>> Pafgbt mailing list
>>>> Pafgbt at listmgr.cv.nrao.edu
>>>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>>
>> _______________________________________________
>> Pafgbt mailing list
>> Pafgbt at listmgr.cv.nrao.edu
>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>>
>
>



More information about the Pafgbt mailing list