[Pafgbt] PAF beamformer size and cost
Scott Ransom
sransom at nrao.edu
Thu Feb 4 14:38:26 EST 2010
For L-band with 500MHz of BW, you'd probably want 2K channels (maybe 1K)
dumping at between 50-100us.
Scott
On Thursday 04 February 2010 02:35:57 pm Rick Fisher wrote:
> John,
>
> How many frequency channels are there and what's the time resolution of
> guppi in search mode?
>
> Rick
>
> On Thu, 4 Feb 2010, John Ford wrote:
> >> Brian,
> >>
> >> Thanks very much for the beamformer outline. How does Jason handle
> >> the frequency dependence of the beamformer weights? Is there an FIR
> >> is each signal path or maybe an FFT - iFFT operation?
> >>
> >> What to do with the high bandwidth beam outputs at high time
> >> resolution is a standard pulsar problem. Since the beam outputs are
> >> essentially the same as you'd have with a conventional horn feed,
> >> all of the current tricks of the trade will apply. As far as I
> >> know, all pulsar surveys are done by generating spectra with a
> >> frequency resolution consistent with the highest expected pulse
> >> dispersion, squaring the signals, integrating for the selected
> >> resolution time, and streaming this to disk. Pulse searches are
> >> then done off-line in computer clusters or supercomputers. (Our
> >> pulsar experts can chime in here.) We'll need to match our output
> >> data rates with available storage and post-processing capabilities -
> >> a time-dependent target. Maybe someone could give us some currently
> >> feasible numbers and time derivatives.
> >
> > I'm not a pulsar expert, but I know a bit about what the 1 beam GBT
> > currently can do, and what is available in the computer world.
> >
> > I don't think it's feasible to keep 40 beams of pulsar search data.
> > We're having a time of it keeping up with managing the data from our
> > 1 beam guppi. We capture ~50-100 MB/sec with guppi in search mode,
> > so 40 times that is completely insane. (20-40 Gigabytes per second!)
> >
> > It's not so much the capture rate (which, while difficult, is
> > possible today with large disk arrays) as the fact that pulsar
> > searches use long integrations, so the volume of data to be stored
> > until it is processed is huge. This example would be >500 Terabytes
> > in an 8 hour observation. Yikes! That's possibly more pulsar search
> > data than has been collected, ever.
> >
> > I have no idea how long it takes to process the data, but the only
> > really feasible thing to do for this kind of data rate is to process
> > it in (near) real time and throw it away when you are done. That
> > means a *really* large supercomputer or special-purpose pulsar search
> > engine.
> >
> > Pulsar timing is another kettle of fish altogether. There, the
> > bottleneck is processing power. (assuming coherent dedispersion) The
> > disk I/O is almost negligible. The I/O requirements for getting the
> > samples off the beamformer and onto the processing machine are also
> > formidable.
> >
> > As usual, the pulsar requirements far outstrip all other observing
> > requirements as far as data rates. We can build spectrometers that
> > integrate for milliseconds or seconds to cut data rates to manageable
> > numbers. So it seems reasonable to build a machine that could handle
> > the 40 beams for "normal" observing, and reduce the number of beams
> > used for pulsar search output.
> >
> > For transients and pulsars you need the fast-sampled time-domain data
> > to work with, and can't really integrate it down much before using
> > it.
> >
> > I think the biggest problem is really storage of the data, and not
> > necessarily the processing in real time.
> >
> > John
> >
> >> Before completely buying into the voltage sum real-time beamformer
> >> we should keep in mind that a lot of single dish applications don't
> >> need voltage outputs as long as the time and frequency resolution
> >> parameters are satisfied. If there are big computational savings in
> >> a post-correlation beamformer, and we satisfy ourselves that there's
> >> not a hidden gotcha in this approach, we should keep it on the
> >> table. My guess is that any computational advantages evaporate or
> >> even reverse when the required time resolution approaches the
> >> inverse required frequency resolution.
> >>
> >> Rick
> >>
> >> On Thu, 4 Feb 2010, Brian Jeffs wrote:
> >>> Rick,
> >>>
> >>> See below:
> >>>> Is your assumed beamformer architecture voltage sums or
> >>>> post-correlation?
> >>>> In other words, are the beams formed by summing complex weighted
> >>>> voltages
> >>>> from the array elements or by combining cross products of all of
> >>>> the elements? John's reference at
> >>>> http://arxiv.org/abs/0912.0380v1 shows a voltage-sum beamformer.
> >>>> The post-correlaion bamformer may use fewer processing resources,
> >>>> but it precludes further coherent signal processing
> >>>> of each beam.
> >>>
> >>> Our plans are based on a correlator/beamformer developed by Jason
> >>> Manley for
> >>> the ATA and some other users (the pocket packetized correlator).
> >>> He recently
> >>> added simultaneous beamforming to the existing correlator gateware,
> >>> so they
> >>> run concurrently. In our application the only time this is
> >>> required is during interference mitigation. Normally we correlate
> >>> during calibration and
> >>> beamform otherwise.
> >>>
> >>> His design is a voltage sum real-time beamformer. At this point he
> >>> does not
> >>> compute as many simultaneous beams as we need to, so I think we
> >>> will have to
> >>> exploit the computational trade-off to do either beamforming or
> >>> correlation
> >>> but not both, or it will not fit in the FPGA. Post-correlation
> >>> beamforming
> >>> is really quite trivial, and has a low computational burden, so
> >>> that could be
> >>> added to the correlator and run simultaneously. I believe that
> >>> when we need
> >>> simultaneous voltage sum beamforming and correlations (as when
> >>> doing interference mitigation) we will have to reduce the effective
> >>> bandwidth. We
> >>> really cannot take Jasons' existing code and plug it right in for
> >>> our application, but it will serve as a very good template. That
> >>> is why we have
> >>> Jonathan out at UC Berkeley for 6 months, so he can learn the ropes
> >>> and then
> >>> work on our correlator/beamformer.
> >>>
> >>>> Very roughly, the science requirements for a beamformer fall into
> >>>> two camps, which may be operational definitions of first science
> >>>> and cadallac/dream machine: 1. spectral line surveys with
> >>>> bandwidths in the 3-100 MHz range and very modest time resolution
> >>>> and 2. pulsar and fast transient source surveys with bandwidths on
> >>>> the order of 500+ MHz and <=50
> >>>> microsecond time resolution. The 2001 science case says pulsar
> >>>> work requires bandwidths of 200+ MHz, but the bar has gone higher
> >>>> in the meantime. One can always think of something to do with a
> >>>> wide bandwidth,
> >>>> low time resolution beamformer, but it would be a stretch. The
> >>>> GBT sensitivity isn't high enough to see HI at redshifts below,
> >>>> say, 1350 MHz
> >>>> in a very wide-area survey. Hence, building a beamformer with
> >>>> wide bandwith but low time resolution may not be the optimum use
> >>>> of resources.
> >>>> Also, the 2001 science cases assumes 7 formed beams, but the
> >>>> minimum now
> >>>> would be, maybe, 19 and growing as the competition heats up.
> >>>
> >>> We are operating under the assumption of at least 19, and probably
> >>> more than
> >>> 40 formed beams. If we only use the correlator for calibration,
> >>> then we should be able to achieve both relatively wide bandwidth
> >>> (250 MHz) and high
> >>> time resolution (we will get a beamformer output per time sample,
> >>> not just
> >>> per STI interval). Dan and Jason feed that based on their
> >>> experience with
> >>> existing codes this is achievable on the 40 ROACH system we
> >>> sketched out, but
> >>> we will have to wait and see. If we run into bottlenecks we will
> >>> have to
> >>> reduce either bandwidth or the number of formed beams.
> >>>
> >>> One issue I am not clear on yet is what we do with the data streams
> >>> for 40+
> >>> voltage sum beams over 500+ frequency channels. How do we get it
> >>> off the
> >>> CASPER array, and what will be done with it? For 8 bit complex
> >>> samples at
> >>> the beamformer outputs you would need the equivalent of fourty 10
> >>> Gbit ethernet links to some other big processor, such as a
> >>> transient detector. If
> >>> this is unreasonable then either the number of bits per sample,
> >>> bandwidth, or
> >>> number of beams will need to be reduced. Alternatively, it is not
> >>> hard to
> >>> add a spectrometer to the beamformer outputs inside the
> >>> correlator/beamformer, and this provides a huge data rate
> >>> reduction. But how
> >>> do we handle data for transient observations where fine time
> >>> resolution is
> >>> critical?
> >>>
> >>> Brian
> >>>
> >>>> Counter-thoughts?
> >>>>
> >>>> Rick
> >>>>
> >>>> On Wed, 3 Feb 2010, Brian Jeffs wrote:
> >>>>> Rick,
> >>>>>
> >>>>> We have a rough architecture and cost estimate for a 40 channel
> >>>>> correlator/beamformer capable of 40 channels (19 dual pol
> >>>>> antennas
> >>>>
> >>>> plus
> >>>>
> >>>>> reference or RFI auxiliary) over 250 MHz BW. We worked this out
> >>>>> with CASOER
> >>>>> head Dan Werthimer and his crack correlator/beamformer developer
> >>>>
> >>>> Jason
> >>>>
> >>>>> Manley. It will require 20 ROACH boards, 20 iADC boards, 1
> >>>>> 20-port
> >>>>
> >>>> 10
> >>>>
> >>>>> Gbit
> >>>>> ethernet switch, and some lesser associated parts.
> >>>>>
> >>>>> Our recent ROACH order was $2750 each, iADC: $1300 each,
> >>>>> enclosures:
> >>>>
> >>>> $750
> >>>>
> >>>>> each, XiLinx chip: free or $3000, ethernet switch: $12000.
> >>>>>
> >>>>> You can use your existing data acquisition array of PCs as the
> >>>>> stream-to-disk
> >>>>> farm, but will need to buy 10 Gbit cards and hardware RAID
> >>>>
> >>>> controllers.
> >>>>
> >>>>> The total (which will be a bit low) assuming no free XiLinx parts
> >>>>> and
> >>>>
> >>>> not
> >>>>
> >>>>> including is: $168,000.
> >>>>>
> >>>>> Of course this does not include development manpower costs.
> >>>>>
> >>>>> Brian
> >>>>>
> >>>>> On Feb 3, 2010, at 3:05 PM, Rick Fisher wrote:
> >>>>>> This is an incomplete question, but maybe we can beat it into
> >>>>
> >>>> something
> >>>>
> >>>>>> answerable: Do we know enough about existing applications on
> >>>>
> >>>> CASPER
> >>>>
> >>>>>> hardware to make a conservative estimate of what it would cost
> >>>>>> to
> >>>>
> >>>> build
> >>>>
> >>>>>> a
> >>>>>> PAF beamformer with a given set of specs? I'm looking for at
> >>>>>> least
> >>>>
> >>>> two
> >>>>
> >>>>>> estimates. What is a realistic set of specs for the first
> >>>>>> science
> >>>>
> >>>> PAF
> >>>>
> >>>>>> beamformer, and what would the dream machine that would make a
> >>>>>> big scientific impact cost? You're welcome to define the specs
> >>>>>> that go with
> >>>>>> either of these two questions or I'll start defining them by
> >>>>
> >>>> thinking
> >>>>
> >>>>>> "out
> >>>>>> loud". The first science beamformer will guide the initial
> >>>>>> system design,
> >>>>>> and the dream machine will help get a handle on longer range
> >>>>>> expectations.
> >>>>>>
> >>>>>> Cheers,
> >>>>>> Rick
> >>>>>>
> >>>>>> _______________________________________________
> >>>>>> Pafgbt mailing list
> >>>>>> Pafgbt at listmgr.cv.nrao.edu
> >>>>>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
> >>>>
> >>>> _______________________________________________
> >>>> Pafgbt mailing list
> >>>> Pafgbt at listmgr.cv.nrao.edu
> >>>> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
> >>
> >> _______________________________________________
> >> Pafgbt mailing list
> >> Pafgbt at listmgr.cv.nrao.edu
> >> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>
> _______________________________________________
> Pafgbt mailing list
> Pafgbt at listmgr.cv.nrao.edu
> http://listmgr.cv.nrao.edu/mailman/listinfo/pafgbt
>
--
Scott M. Ransom Address: NRAO
Phone: (434) 296-0320 520 Edgemont Rd.
email: sransom at nrao.edu Charlottesville, VA 22903 USA
GPG Fingerprint: 06A9 9553 78BE 16DB 407B FFCA 9BFA B6FF FFD3 2989
More information about the Pafgbt
mailing list