Glad you were able to figure it out. FWIW, I had the same interpretation as you. Let us know if you need anything else.
- Prateek On Tue, Apr 2, 2019 at 4:55 PM Malcolm McFarland <mmcfarl...@cavulus.com> wrote: > Found the issue, and thank goodness it was a configuration issue on my end: > I was setting the yarn.scheduler.maximum-allocation-vcores too low and > artificially constraining the cluster. This is (as the name implies) the > maximum allocation for the entire cluster; I had interpreted the > description from the v2.6.1 docs (which I was initially using because of > its inclusion in the hello-samza project) to mean that this was a > per-container setting. > > Thanks again for the help, and for the tip on upgrading to Yarn 2.7.6! > > Cheers, > Malcolm > > On Tue, Apr 2, 2019 at 1:47 PM Malcolm McFarland <mmcfarl...@cavulus.com> > wrote: > > > Interestingly, I just tried setting > > yarn.scheduler.minimum-allocation-vcores=2 and restarting everything. On > > startup, the RM now displays a Minimum Allocation of <memory:256, > > vCores:2>, but my application container still shows "Resource:4096 > Memory, > > 1 VCores". The statistics page for the "default" queue shows "Used > > Resources:<memory:13312, vCores:6>...Num Containers:6", which is accurate > > (3 tasks + 3 AMs). > > > > This seems a long shot, but is there a chance that I'm reading this > > incorrectly, and that YARN will show CPU usage when the processes > actually > > start processing -- ie, is the resource allocation shown on-demand, as > > opposed to preemptive? > > > > Cheers, > > Malcolm > > > > Cheers, > > Malcolm > > > > > > On Tue, Apr 2, 2019 at 12:54 PM Malcolm McFarland < > mmcfarl...@cavulus.com> > > wrote: > > > >> Hi Prateek, > >> > >> I'm not getting an error now, just an unyielding vcore allotment of 1. > >> I just verified that we're setting > >> > >> > yarn.resourcemanager.scheduler.class=org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler > >> and > >> > yarn.scheduler.capacity.resource-calculator=org.apache.hadoop.yarn.util.resource.DominantResourceCalculator > >> via the configuration tab in the RM UI. (Interestingly, although the > >> resource-calculator property is discussed on the page to which you > >> linked, it's not on the yarn-defaults.xml reference page for v2.7.6). > >> Pretty much every configuration option related to max-vcores is set to > >> a number >1, as is the cluster-manager.container.cpu.cores setting in > >> Samza. So although no errors, still just a single core per container. > >> > >> Here's a question: in Samza, are the cluster resource allocation > >> properties pulled from a properties file in the application bundle, or > >> are they sourced from the properties file that is passed to run-app.sh > >> (and used to submit the task to YARN)? Are there any other properties > >> for which this would make a difference? > >> > >> Cheers, > >> Malcolm > >> > >> > >> On Tue, Apr 2, 2019 at 9:50 AM Prateek Maheshwari <prateek...@gmail.com > > > >> wrote: > >> > > >> > And just to double check, you also changed the > >> > yarn.resourcemanager.scheduler.class to CapacityScheduler? > >> > > >> > On Tue, Apr 2, 2019 at 9:49 AM Prateek Maheshwari < > prateek...@gmail.com > >> > > >> > wrote: > >> > > >> > > Is it still the same message from the AM? The one that says: "Got AM > >> > > register response. The YARN RM supports container requests with > >> max-mem: > >> > > 14336, max-cpu: 1" > >> > > > >> > > On Tue, Apr 2, 2019 at 12:09 AM Malcolm McFarland < > >> mmcfarl...@cavulus.com> > >> > > wrote: > >> > > > >> > >> Hey Prateek, > >> > >> > >> > >> The upgrade to Hadoop 2.7.6 went fine; everything seems to be > >> working, and > >> > >> access to S3 via an access key/secret pair is working as well. > >> However, my > >> > >> requested tasks are still only getting allocated 1 core, despite > >> > >> requesting > >> > >> more than that. Once again, I have a 3-node cluster that should > have > >> 24 > >> > >> vcores available; on the yarn side, I have these options set: > >> > >> > >> > >> nodemanager.resource.cpu-vcores=8 > >> > >> yarn.scheduler.minimum-allocation-vcore=1 > >> > >> yarn.scheduler.maximum-allocation-vcores=4 > >> > >> > >> > >> > >> > yarn.scheduler.capacity.resource-calculator=org.apache.hadoop.yarn.util.resource.DominantResourceCalculator > >> > >> > >> > >> And on the Samza side, I'm setting: > >> > >> > >> > >> cluster-manager.container.cpu.cores=2 > >> > >> > >> > >> However, YARN is still telling me that the running task has 1 vcore > >> > >> assigned. Do you have any other suggestions for options to tweak? > >> > >> > >> > >> Cheers, > >> > >> Malcolm > >> > >> > >> > >> > >> > >> On Mon, Apr 1, 2019 at 5:28 PM Malcolm McFarland < > >> mmcfarl...@cavulus.com> > >> > >> wrote: > >> > >> > >> > >> > One more thing -- fwiw, I actually also came across the > possibility > >> > >> that I > >> > >> > would need to use the DominantResourceCalculator, but as you > point > >> out, > >> > >> > this doesn't seem to be available in Hadoop 2.6. > >> > >> > > >> > >> > > >> > >> > On Mon, Apr 1, 2019 at 5:27 PM Malcolm McFarland < > >> > >> mmcfarl...@cavulus.com> > >> > >> > wrote: > >> > >> > > >> > >> >> That's quite helpful! I actually initially tried using a version > >> of > >> > >> >> Hadoop > 2.6.x; when I did, it seemed like the AWS credentials > in > >> YARN > >> > >> >> (fs.s3a.access.key, fs.s3a.secret.key) weren't being accessed, > as > >> I > >> > >> >> received lots of "No AWS Credentials > >> > >> >> provided by DefaultAWSCredentialsProviderChain" messages. I > found > >> a > >> > >> >> way around this by providing the credentials to the AM directly > >> via > >> > >> >> yarn.am.opts=-Daws.accessKeyId=<key> -Daws.secretKey=<secret>, > but > >> > >> >> since this seemed very workaround-ish, I just assumed that I > would > >> > >> >> eventually hit other problems using a version of Hadoop not > >> pinned in > >> > >> >> the Samza repo. If you're running 2.7.x at LinkedIn, however, > I'll > >> > >> >> give it a shot again. > >> > >> >> > >> > >> >> Have you done any AWS credential integration, and if so, did you > >> need > >> > >> >> to do anything special to get it to work? > >> > >> >> > >> > >> >> Cheers, > >> > >> >> Malcolm > >> > >> >> > >> > >> >> > >> > >> >> > >> > >> >> On Mon, Apr 1, 2019 at 5:20 PM Prateek Maheshwari < > >> > >> prateek...@gmail.com> > >> > >> >> wrote: > >> > >> >> > > >> > >> >> > Hi Malcolm, > >> > >> >> > > >> > >> >> > I think this is because in YARN 2.6 the FifoScheduler only > >> accounts > >> > >> for > >> > >> >> > memory for 'maximumAllocation': > >> > >> >> > > >> > >> >> > >> > >> > >> > https://github.com/apache/hadoop/blob/branch-2.6.2/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java#L218 > >> > >> >> > > >> > >> >> > This has been changed as early as 2.7.0: > >> > >> >> > > >> > >> >> > >> > >> > >> > https://github.com/apache/hadoop/blob/branch-2.7.0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java#L218 > >> > >> >> > > >> > >> >> > So upgrading will likely fix this issue. For reference, at > >> LinkedIn > >> > >> we > >> > >> >> are > >> > >> >> > running YARN 2.7.2 with the CapacityScheduler > >> > >> >> > < > >> > >> >> > >> > >> > >> > https://hadoop.apache.org/docs/r2.7.2/hadoop-yarn/hadoop-yarn-site/CapacityScheduler.html > >> > >> >> > > >> > >> >> > and DominantResourceCalculator to account for vcore > allocations > >> in > >> > >> >> > scheduling. > >> > >> >> > > >> > >> >> > - Prateek > >> > >> >> > > >> > >> >> > On Mon, Apr 1, 2019 at 3:00 PM Malcolm McFarland < > >> > >> >> mmcfarl...@cavulus.com> > >> > >> >> > wrote: > >> > >> >> > > >> > >> >> > > Hi Prateek, > >> > >> >> > > > >> > >> >> > > This still seems to be manifesting with the same problem. > >> Since > >> > >> this > >> > >> >> seems > >> > >> >> > > to be something in the hadoop codebase, and I've emailed the > >> > >> >> hadoop-dev > >> > >> >> > > mailing list about it. > >> > >> >> > > > >> > >> >> > > Cheers, > >> > >> >> > > Malcolm > >> > >> >> > > > >> > >> >> > > On Mon, Apr 1, 2019 at 1:51 PM Prateek Maheshwari < > >> > >> >> prateek...@gmail.com> > >> > >> >> > > wrote: > >> > >> >> > > > >> > >> >> > > > Hi Malcolm, > >> > >> >> > > > > >> > >> >> > > > Yes, the AM is just reporting what the RM specified as the > >> > >> maximum > >> > >> >> > > allowed > >> > >> >> > > > request size. > >> > >> >> > > > > >> > >> >> > > > I think 'yarn.scheduler.maximum-allocation-vcores' needs > to > >> be > >> > >> less > >> > >> >> than > >> > >> >> > > > 'yarn.nodemanager.resource.cpu-vcores', since a container > >> must > >> > >> fit > >> > >> >> on a > >> > >> >> > > > single NM. Maybe the RM detected this and decided to > >> default to > >> > >> 1? > >> > >> >> Can > >> > >> >> > > you > >> > >> >> > > > try setting maximum-allocation-vcores lower? > >> > >> >> > > > > >> > >> >> > > > - Prateek > >> > >> >> > > > > >> > >> >> > > > On Mon, Apr 1, 2019 at 11:59 AM Malcolm McFarland < > >> > >> >> > > mmcfarl...@cavulus.com> > >> > >> >> > > > wrote: > >> > >> >> > > > > >> > >> >> > > > > One other detail: I'm running YARN on ECS in AWS. Has > >> anybody > >> > >> seen > >> > >> >> > > > > issues with core allocation in this environment? I'm > >> seeing > >> > >> this > >> > >> >> in > >> > >> >> > > > > the samza log: > >> > >> >> > > > > > >> > >> >> > > > > "Got AM register response. The YARN RM supports > container > >> > >> requests > >> > >> >> > > > > with max-mem: 14336, max-cpu: 1" > >> > >> >> > > > > > >> > >> >> > > > > How does samza determine this? Looking at the Samza > >> source on > >> > >> >> Github, > >> > >> >> > > > > it appears to be information that's passed back to the > AM > >> when > >> > >> it > >> > >> >> > > > > starts up. > >> > >> >> > > > > > >> > >> >> > > > > Cheers, > >> > >> >> > > > > Malcolm > >> > >> >> > > > > > >> > >> >> > > > > On Mon, Apr 1, 2019 at 10:44 AM Malcolm McFarland > >> > >> >> > > > > <mmcfarl...@cavulus.com> wrote: > >> > >> >> > > > > > > >> > >> >> > > > > > Hi Prateek, > >> > >> >> > > > > > > >> > >> >> > > > > > Sorry, meant to include these versions with my email; > >> I'm > >> > >> >> running > >> > >> >> > > > > > Samza 0.14 and Hadoop 2.6.1. I'm running three > >> containers > >> > >> >> across 3 > >> > >> >> > > > > > node managers, each with 16GB and 8 vcores. The other > >> two > >> > >> >> containers > >> > >> >> > > > > > are requesting 1 vcore each; even with the AMs > running, > >> that > >> > >> >> should > >> > >> >> > > be > >> > >> >> > > > > > 4 for them in total, leaving plenty of processing > power > >> > >> >> available. > >> > >> >> > > > > > > >> > >> >> > > > > > The error is in the application attempt diagnostics > >> field: > >> > >> "The > >> > >> >> YARN > >> > >> >> > > > > > cluster is unable to run your job due to unsatisfiable > >> > >> resource > >> > >> >> > > > > > requirements. You asked for mem: 2048, and cpu: 2." I > >> do not > >> > >> >> see this > >> > >> >> > > > > > error with the same memory request, but a cpu count > >> request > >> > >> of > >> > >> >> 1. > >> > >> >> > > > > > > >> > >> >> > > > > > Here are the configuration options pertaining to > >> resource > >> > >> >> allocation: > >> > >> >> > > > > > > >> > >> >> > > > > > <?xml version="1.0"?> > >> > >> >> > > > > > <configuration> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.resourcemanager.scheduler.class</name> > >> > >> >> > > > > > > >> > >> >> > > > > > >> > >> >> > > > > >> > >> >> > > > >> > >> >> > >> > >> > >> > <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.nodemanager.vmem-check-enabled</name> > >> > >> >> > > > > > <value>false</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.nodemanager.vmem-pmem-ratio</name> > >> > >> >> > > > > > <value>2.1</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.nodemanager.resource.memory-mb</name> > >> > >> >> > > > > > <value>14336</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.scheduler.minimum-allocation-mb</name> > >> > >> >> > > > > > <value>256</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.scheduler.maximum-allocation-mb</name> > >> > >> >> > > > > > <value>14336</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > > >> <name>yarn.scheduler.minimum-allocation-vcores</name> > >> > >> >> > > > > > <value>1</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > > >> <name>yarn.scheduler.maximum-allocation-vcores</name> > >> > >> >> > > > > > <value>16</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.nodemanager.resource.cpu-vcores</name> > >> > >> >> > > > > > <value>8</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > <property> > >> > >> >> > > > > > <name>yarn.resourcemanager.cluster-id</name> > >> > >> >> > > > > > <value>processor-cluster</value> > >> > >> >> > > > > > </property> > >> > >> >> > > > > > </configuration> > >> > >> >> > > > > > > >> > >> >> > > > > > Cheers, > >> > >> >> > > > > > Malcolm > >> > >> >> > > > > > > >> > >> >> > > > > > On Mon, Apr 1, 2019 at 10:25 AM Prateek Maheshwari < > >> > >> >> > > > prateek...@gmail.com> > >> > >> >> > > > > wrote: > >> > >> >> > > > > > > > >> > >> >> > > > > > > Hi Malcolm, > >> > >> >> > > > > > > > >> > >> >> > > > > > > Just setting that configuration should be > sufficient. > >> We > >> > >> >> haven't > >> > >> >> > > seen > >> > >> >> > > > > this > >> > >> >> > > > > > > issue before. What Samza/YARN versions are you > using? > >> Can > >> > >> you > >> > >> >> also > >> > >> >> > > > > include > >> > >> >> > > > > > > the logs from where you get the error and your yarn > >> > >> >> configuration? > >> > >> >> > > > > > > > >> > >> >> > > > > > > - Prateek > >> > >> >> > > > > > > > >> > >> >> > > > > > > On Mon, Apr 1, 2019 at 2:33 AM Malcolm McFarland < > >> > >> >> > > > > mmcfarl...@cavulus.com> > >> > >> >> > > > > > > wrote: > >> > >> >> > > > > > > > >> > >> >> > > > > > > > Hey Folks, > >> > >> >> > > > > > > > > >> > >> >> > > > > > > > I'm having some issues getting multiple cores for > >> > >> >> containers in > >> > >> >> > > > yarn. > >> > >> >> > > > > > > > I seem to have my YARN settings correct, and the > RM > >> > >> >> interface > >> > >> >> > > says > >> > >> >> > > > > > > > that I have 24vcores available. However, when I > set > >> the > >> > >> >> > > > > > > > cluster-manager.container.cpu.cores Samza setting > to > >> > >> >> anything > >> > >> >> > > other > >> > >> >> > > > > > > > than 1, I get a message about how the container is > >> > >> >> requesting > >> > >> >> > > more > >> > >> >> > > > > > > > resources than it can allocate. With 1 core, > >> everything > >> > >> is > >> > >> >> fine. > >> > >> >> > > Is > >> > >> >> > > > > > > > there another Samza option I need to set? > >> > >> >> > > > > > > > > >> > >> >> > > > > > > > Cheers, > >> > >> >> > > > > > > > Malcolm > >> > >> >> > > > > > > > > >> > >> >> > > > > > > > > >> > >> >> > > > > > > > -- > >> > >> >> > > > > > > > Malcolm McFarland > >> > >> >> > > > > > > > Cavulus > >> > >> >> > > > > > > > > >> > >> >> > > > > > > >> > >> >> > > > > > > >> > >> >> > > > > > > >> > >> >> > > > > > -- > >> > >> >> > > > > > Malcolm McFarland > >> > >> >> > > > > > Cavulus > >> > >> >> > > > > > 1-800-760-6915 > >> > >> >> > > > > > mmcfarl...@cavulus.com > >> > >> >> > > > > > > >> > >> >> > > > > > > >> > >> >> > > > > > This correspondence is from HealthPlanCRM, LLC, d/b/a > >> > >> Cavulus. > >> > >> >> Any > >> > >> >> > > > > > unauthorized or improper disclosure, copying, > >> distribution, > >> > >> or > >> > >> >> use of > >> > >> >> > > > > > the contents of this message is prohibited. The > >> information > >> > >> >> contained > >> > >> >> > > > > > in this message is intended only for the personal and > >> > >> >> confidential > >> > >> >> > > use > >> > >> >> > > > > > of the recipient(s) named above. If you have received > >> this > >> > >> >> message in > >> > >> >> > > > > > error, please notify the sender immediately and delete > >> the > >> > >> >> original > >> > >> >> > > > > > message. > >> > >> >> > > > > > >> > >> >> > > > > > >> > >> >> > > > > > >> > >> >> > > > > -- > >> > >> >> > > > > Malcolm McFarland > >> > >> >> > > > > Cavulus > >> > >> >> > > > > 1-800-760-6915 > >> > >> >> > > > > mmcfarl...@cavulus.com > >> > >> >> > > > > > >> > >> >> > > > > > >> > >> >> > > > > This correspondence is from HealthPlanCRM, LLC, d/b/a > >> Cavulus. > >> > >> Any > >> > >> >> > > > > unauthorized or improper disclosure, copying, > >> distribution, or > >> > >> >> use of > >> > >> >> > > > > the contents of this message is prohibited. The > >> information > >> > >> >> contained > >> > >> >> > > > > in this message is intended only for the personal and > >> > >> >> confidential use > >> > >> >> > > > > of the recipient(s) named above. If you have received > this > >> > >> >> message in > >> > >> >> > > > > error, please notify the sender immediately and delete > the > >> > >> >> original > >> > >> >> > > > > message. > >> > >> >> > > > > > >> > >> >> > > > > >> > >> >> > > > >> > >> >> > > > >> > >> >> > > -- > >> > >> >> > > Malcolm McFarland > >> > >> >> > > Cavulus > >> > >> >> > > 1-800-760-6915 > >> > >> >> > > mmcfarl...@cavulus.com > >> > >> >> > > > >> > >> >> > > > >> > >> >> > > This correspondence is from HealthPlanCRM, LLC, d/b/a > >> Cavulus. Any > >> > >> >> > > unauthorized or improper disclosure, copying, distribution, > >> or use > >> > >> of > >> > >> >> the > >> > >> >> > > contents of this message is prohibited. The information > >> contained > >> > >> in > >> > >> >> this > >> > >> >> > > message is intended only for the personal and confidential > >> use of > >> > >> the > >> > >> >> > > recipient(s) named above. If you have received this message > in > >> > >> error, > >> > >> >> > > please notify the sender immediately and delete the original > >> > >> message. > >> > >> >> > > > >> > >> >> > >> > >> >> > >> > >> >> > >> > >> >> -- > >> > >> >> Malcolm McFarland > >> > >> >> Cavulus > >> > >> >> 1-800-760-6915 > >> > >> >> mmcfarl...@cavulus.com > >> > >> >> > >> > >> >> > >> > >> >> This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. > Any > >> > >> >> unauthorized or improper disclosure, copying, distribution, or > >> use of > >> > >> >> the contents of this message is prohibited. The information > >> contained > >> > >> >> in this message is intended only for the personal and > >> confidential use > >> > >> >> of the recipient(s) named above. If you have received this > >> message in > >> > >> >> error, please notify the sender immediately and delete the > >> original > >> > >> >> message. > >> > >> >> > >> > >> > > >> > >> > > >> > >> > -- > >> > >> > Malcolm McFarland > >> > >> > Cavulus > >> > >> > 1-800-760-6915 > >> > >> > mmcfarl...@cavulus.com > >> > >> > > >> > >> > > >> > >> > This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. > Any > >> > >> > unauthorized or improper disclosure, copying, distribution, or > use > >> of > >> > >> the > >> > >> > contents of this message is prohibited. The information contained > >> in > >> > >> this > >> > >> > message is intended only for the personal and confidential use of > >> the > >> > >> > recipient(s) named above. If you have received this message in > >> error, > >> > >> > please notify the sender immediately and delete the original > >> message. > >> > >> > > >> > >> > >> > >> > >> > >> -- > >> > >> Malcolm McFarland > >> > >> Cavulus > >> > >> 1-800-760-6915 > >> > >> mmcfarl...@cavulus.com > >> > >> > >> > >> > >> > >> This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. Any > >> > >> unauthorized or improper disclosure, copying, distribution, or use > >> of the > >> > >> contents of this message is prohibited. The information contained > in > >> this > >> > >> message is intended only for the personal and confidential use of > the > >> > >> recipient(s) named above. If you have received this message in > error, > >> > >> please notify the sender immediately and delete the original > message. > >> > >> > >> > > > >> > >> > >> > >> -- > >> Malcolm McFarland > >> Cavulus > >> 1-800-760-6915 > >> mmcfarl...@cavulus.com > >> > >> > >> This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. Any > >> unauthorized or improper disclosure, copying, distribution, or use of > >> the contents of this message is prohibited. The information contained > >> in this message is intended only for the personal and confidential use > >> of the recipient(s) named above. If you have received this message in > >> error, please notify the sender immediately and delete the original > >> message. > >> > > > > > > -- > > Malcolm McFarland > > Cavulus > > 1-800-760-6915 > > mmcfarl...@cavulus.com > > > > > > This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. Any > > unauthorized or improper disclosure, copying, distribution, or use of the > > contents of this message is prohibited. The information contained in this > > message is intended only for the personal and confidential use of the > > recipient(s) named above. If you have received this message in error, > > please notify the sender immediately and delete the original message. > > > > > -- > Malcolm McFarland > Cavulus > 1-800-760-6915 > mmcfarl...@cavulus.com > > > This correspondence is from HealthPlanCRM, LLC, d/b/a Cavulus. Any > unauthorized or improper disclosure, copying, distribution, or use of the > contents of this message is prohibited. The information contained in this > message is intended only for the personal and confidential use of the > recipient(s) named above. If you have received this message in error, > please notify the sender immediately and delete the original message. >