NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
I see two options (since you don't want to use two machines). If you insert the products into the queue originally, then you could insert each set of data as a seperate feedtype. If this is an existing feed, that you need to split up, then you could put a matching pattern in your pqact.conf and reinsert the product into the queue using pqinsert and specifying a different feedtype. Then just allow by feedtypes as usual. Should work, haven't tried it. David > It's really that they SHOULDN'T get any data that is not intended to them. > > Daniel > > > -----Original Message----- > > From: Steve Emmerson [mailto:steve@xxxxxxxxxxxxxxxx] > > Sent: 21 November, 2003 17:07 > > To: Lemay,Daniel [CMC] > > Cc: ldm-users@xxxxxxxxxxxxxxxx > > Subject: Re: Many instances of LDM on the same machine > > > > > > Daniel, > > > > >Date: Fri, 21 Nov 2003 16:45:47 -0500 > > >From: "Lemay,Daniel [CMC]" <Daniel.Lemay@xxxxxxxx> > > >To: "'Steve Emmerson'" <steve@xxxxxxxxxxxxxxxx> > > >Subject: RE: Many instances of LDM on the same machine > > > > The above message contained the following: > > > > > Suppose that I'm in the following situation: > > ... > > > -I want to be sure that my "clients" can only obtain the files that > > > are intend for them. I don't want to rely on the fact they > > will choose > > > themselves the subset of the feed I allow them. I know that > > I cannot > > > insert a regular expression in my ALLOW statement to restrict what > > > files they will have access. > > > > That's an interesting situation (and one that I haven't considered). > > > > Is it the case that each downstream site SHOULDN'T get the > > other's data or that such feed-requests would merely be inefficient? > > > > Regards, > > Steve Emmerson > >
ldm-users
archives: