I am using this job design that involves a CASS stage. The max through put i get is 1500 rows/sec. Is there a way to make it tun better??
DS--->Xmr-->CASS-->DS.
what i tried to improve the performnace.
1. played with bumping up the config from 2,4, then 6. 2 to 4 made a little difference after that no difference
2. Partition the data and include multiple CASS stages like below and no differnce.
DS --------xmr------> CASSS ----> FUNNEL --->DS
split the data|------> CASS----->
split the data|------->CASS ------>
split the data|-------> CASS------>
does the installation location of CASS DB makes any difference? I see it is installed on the same mount where IIS is installed.
can you share your experience..I really appreciate any inputs on this.
Thanks.
CASS stage performance
-
- Participant
- Posts: 152
- Joined: Tue Jan 13, 2009 8:59 am
Re: CASS stage performance
update: I tried the below also
1. Made the CASS db installed on multiple mounts and made multiple CASS stages with in the same job and still the same.
2. Tried to do the same in another GRID env making the job to run only on the conductor node still no use
at this point we raised a PMR with IBM and see what they have to say.
The challenge is the fact that we don't have a general ball park of how efficiently this stage runs. I really appreciate your experiences using CASS stage.
1. Made the CASS db installed on multiple mounts and made multiple CASS stages with in the same job and still the same.
2. Tried to do the same in another GRID env making the job to run only on the conductor node still no use
at this point we raised a PMR with IBM and see what they have to say.
The challenge is the fact that we don't have a general ball park of how efficiently this stage runs. I really appreciate your experiences using CASS stage.
-
- Participant
- Posts: 152
- Joined: Tue Jan 13, 2009 8:59 am
CASS stage performance
Coming back here to update:
we had a call with IBM and they mentioned that CASS scales with number of CPUs and very CPU intensive, so we went back and tried with a 8 cpu config file and there was no difference when compared with 4 cpu. The reason is that with 4Node config file(our hardware is 4 cpu in this case), when we issue NMON the cpus are already clocking close to 95-97% and so CASS wasn't scaling after 4 Node config file.
To verify the scalability, we ran CASS in an environment with 24 CPU and ran with 4,8 and up 16 Node config files and we saw CASS scaling with increase in number of Nodes until the CPUs max out. we saw upto 7k/sec with 16 cpu.
I hope this helps. btw, in IBM lab testing they mentioned that expecting 200rows/sec per cpu is the average benchmark. It may vary with each individual environment.
@rjdickson, what difference does it make if we do hash and sort on ZIP? i am curious. It hasn't made any difference though.
we had a call with IBM and they mentioned that CASS scales with number of CPUs and very CPU intensive, so we went back and tried with a 8 cpu config file and there was no difference when compared with 4 cpu. The reason is that with 4Node config file(our hardware is 4 cpu in this case), when we issue NMON the cpus are already clocking close to 95-97% and so CASS wasn't scaling after 4 Node config file.
To verify the scalability, we ran CASS in an environment with 24 CPU and ran with 4,8 and up 16 Node config files and we saw CASS scaling with increase in number of Nodes until the CPUs max out. we saw upto 7k/sec with 16 cpu.
I hope this helps. btw, in IBM lab testing they mentioned that expecting 200rows/sec per cpu is the average benchmark. It may vary with each individual environment.
@rjdickson, what difference does it make if we do hash and sort on ZIP? i am curious. It hasn't made any difference though.
Thanks for the update! Please take the 200 rows/sec average with a grain of salt. For comparison:
On 4 CPUs, IBM Power7 hardware likely dated 2011 or 2012, with a 4 node config file, we get nearly 300 rows/sec on a bad day, and 2600 rows/sec on a good day (over 10x the "average"). This is using the same input file to compare against a baseline. We repeat this same test at least every time the reference files get updated, month after month, year after year.
Our performance variation is consistently inconsistent, meaning it is either fast or slow, with nothing ever in between. I have not spent any time yet to dive in and examine the job or the server. My burning question is why the extreme variation?
I would of course prefer to get the faster performance every time. I can imagine maybe SAN disk caching making the difference but it would some time and effort to prove that one way or another.
On 4 CPUs, IBM Power7 hardware likely dated 2011 or 2012, with a 4 node config file, we get nearly 300 rows/sec on a bad day, and 2600 rows/sec on a good day (over 10x the "average"). This is using the same input file to compare against a baseline. We repeat this same test at least every time the reference files get updated, month after month, year after year.
Our performance variation is consistently inconsistent, meaning it is either fast or slow, with nothing ever in between. I have not spent any time yet to dive in and examine the job or the server. My burning question is why the extreme variation?
I would of course prefer to get the faster performance every time. I can imagine maybe SAN disk caching making the difference but it would some time and effort to prove that one way or another.
Choose a job you love, and you will never have to work a day in your life. - Confucius