Thanks tim. Is my assumption is right ?
Throughput depends on CPU threads and sockets arch On Friday, September 22, 2017 at 12:39:41 PM UTC+5:30, Tim Hockin wrote: > > Thanks for following up! > > On Thu, Sep 21, 2017 at 1:33 AM, Vinoth Narasimhan <talk2...@gmail.com > <javascript:>> wrote: > > Finally the issue was with the hardware spec. The previous k8s test i > did > > with 3 node cluster with each node spec has 1cpu and 4Gig RAM. > > > > Today i map the spec of the native tomcat test machine with the K8s > node. > > > > I created the new 3-node K8s cluster with version 1.6.9 with node > > configuration 8CPU and 30Gig RAM each. > > > > In this setup i see a consistent 8k Request/handling both in POD-POD > > communication as well as from Loadbalancer Endpoints from K8s. > > > > > > In the resource usage none of the cluster hitting the limit.CPU is well > with > > at max less than 13% usage on both the cluster. > > > > When i see the CPU information from > > lscpu > > > > This is the CPU sec for the K8s node where we get the 2k req/sec numbers > > > > Architecture: x86_64 > > CPU op-mode(s): 32-bit, 64-bit > > Byte Order: Little Endian > > CPU(s): 1 > > On-line CPU(s) list: 0 > > Thread(s) per core: 1 > > Core(s) per socket: 1 > > Socket(s): 1 > > NUMA node(s): 1 > > Vendor ID: GenuineIntel > > CPU family: 6 > > Model: 62 > > Model name: Intel(R) Xeon(R) CPU @ 2.50GHz > > Stepping: 4 > > CPU MHz: 2500.000 > > BogoMIPS: 5000.00 > > Hypervisor vendor: KVM > > Virtualization type: full > > L1d cache: 32K > > L1i cache: 32K > > L2 cache: 256K > > L3 cache: 30720K > > NUMA node0 CPU(s): 0 > > Flags > > > > > > > > =================================================================== > > > > This is the CPU sec for the K8s node where we get the 8k req/sec numbers > > > > > > Architecture: x86_64 > > CPU op-mode(s): 32-bit, 64-bit > > Byte Order: Little Endian > > CPU(s): 8 > > On-line CPU(s) list: 0-7 > > Thread(s) per core: 2 > > Core(s) per socket: 4 > > Socket(s): 1 > > NUMA node(s): 1 > > Vendor ID: GenuineIntel > > CPU family: 6 > > Model: 62 > > Model name: Intel(R) Xeon(R) CPU @ 2.50GHz > > Stepping: 4 > > CPU MHz: 2500.000 > > BogoMIPS: 5000.00 > > Hypervisor vendor: KVM > > Virtualization type: full > > L1d cache: 32K > > L1i cache: 32K > > L2 cache: 256K > > L3 cache: 30720K > > NUMA node0 CPU(s): 0-7 > > Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr > pge > > mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb > rdtscp > > lm constant_tsc rep_good nopl xtopology nonstop_tsc eagerfpu pni > pclmulqdq > > ssse3 cx16 sse4_1 sse4_2 x2apic popcnt aes xsave avx f16c rdrand > hypervisor > > lahf_lm fsgsbase tsc_adjust smep erms xsaveopt > > > > > > > > > > > > I think the throughput all depends on the following CPU architecture i > > guess. > > > > > > > > Thread(s) per core:2,Core(s) per socket:4 > > > > Thanks for all your valuable suggestions. > > > > On Tuesday, September 19, 2017 at 11:55:05 AM UTC+5:30, Vinoth > Narasimhan > > wrote: > >> > >> Environment: > >> > >> Kubernetes version (use kubectl version): > >> kubectl version > >> Client Version: version.Info{Major:"1", Minor:"7", GitVersion:"v1.7.3", > >> GitCommit:"2c2fe6e8278a5db2d15a013987b53968c743f2a1", > GitTreeState:"clean", > >> BuildDate:"2017-08-03T07:00:21Z", GoVersion:"go1.8.3", Compiler:"gc", > >> Platform:"linux/amd64"} > >> Server Version: version.Info{Major:"1", Minor:"6", GitVersion:"v1.6.9", > >> GitCommit:"a3d1dfa6f433575ce50522888a0409397b294c7b", > GitTreeState:"clean", > >> BuildDate:"2017-08-23T16:58:45Z", GoVersion:"go1.7.6", Compiler:"gc", > >> Platform:"linux/amd64"} > >> > >> Cloud provider or hardware configuration**: > >> > >> Google Container Engine. > >> > >> What happened: > >> > >> We are in testing phase of springboot based microservice deployment on > >> GKE. During testing QA filed a performance issue , stats that the > throughput > >> of the service in k8s is low when compared to run the java app in > >> > >> java -jar method > >> docker run > >> For testing i skip those springboot stuff and take native tomcat home > page > >> as the test bed for the "ab" testing. > >> > >> The test setup looks like: > >> > >> Create an 8cpu/30Gig RAM ubuntu server in GCP and install native > >> tomact-8.5.20(80) and test the home page. > >> > >> Stop the native tomcat. Create the docker tomcat instances on the same > >> host and test the same home page. > >> The docker version is: Version: 17.06.2-ce > >> > >> Create the 3 node K8s cluster 1.6.9. Run the tomcat deployment the same > >> 8.5.20 and expose the service through LB and test the same home page. > >> > >> I install the ab tool in other GCP instances and hit the above 3 > different > >> endpoints. > >> > >> What's the Result: > >> > >> The first 2 test with native tomcat and docker run the throughput i got > is > >> nearly 8k Req/sec on avg on different request/concurrent level. > >> > >> But the same on K8s LB the throughput i got on the average of 2k > req/sec > >> on avg on different request/concurrency level. > >> > >> Is this something am i missing on the test. Or this is how the GKE LB > >> store and forward the request at this rate. > > > > -- > > You received this message because you are subscribed to the Google > Groups > > "Kubernetes user discussion and Q&A" group. > > To unsubscribe from this group and stop receiving emails from it, send > an > > email to kubernetes-use...@googlegroups.com <javascript:>. > > To post to this group, send email to kubernet...@googlegroups.com > <javascript:>. > > Visit this group at https://groups.google.com/group/kubernetes-users. > > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "Kubernetes user discussion and Q&A" group. To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-users+unsubscr...@googlegroups.com. To post to this group, send email to kubernetes-users@googlegroups.com. Visit this group at https://groups.google.com/group/kubernetes-users. For more options, visit https://groups.google.com/d/optout.