I think this may also be a problem resulting from recent 2.0 changes.
So, everything is compiling and will run.
However, I’m configured to share connections - because I’m not really simulating unique users. And I’m configured to use a max number of connections*
Right now I’m testing against our servers heartbeat service - the service used by our ELB to determine whether the service is there and responsive.
I’m getting responses back, and everything OKs …
… until I’ve reached the number of requests equal to the number of max connections.
So, for example, I have 10K total connections. I make 500 requests per second. After 20s, I begin getting exception spam from not having enough connections to run the new user simulations.
This use to work before. It was my belief that once I had an OK, that the connection that had been used to generate the OK was freed for a new user.
It doesn’t appear to be working that way right now. Is there a new setting that was introduced? Can someone confirm that they aren’t seeing the same thing on their machine?
I’ll follow-up with a stripped down example of what I’m doing.
- When I began using gatling last year, I realized that as latency increased more connections would get spawned to backlog users. So I could never get a real honest sense of throughput and server performance. I began to use max connections so that when that number of connections was exceeded, I would realize there was some kind of performance problem - either the server was underperforming or there were some kind of latency hiccups or I needed more connections to accomodate the latency. It gives me a better sense of what kind of performance I’m really talking about. E.g. our server can handle 10K connections, but it can only handle 6K requests through that pool of connections. I pair these observations with CPU performance and memory usage on the server.
How long are your requests?
I mean that if you send 500 req/sec but your requests last 3 sec, you might need 1500 connections.
Then, is it AsyncHttpClient that complains, or the OS (file descriptors starvation)?
The requests are all < 500ms
The running requests is always reported as < 50, with completed requests registering as OK
Attached, you’ll find my pom (named run.xml, with my jvm args), my gatling.conf, and my Demo class which recreates identically the problem I’m having.
It looks like it’s akka that’s complaining. The stack trace that prints out once I run out of connections is:
[ERROR] [03/31/2014 14:53:28.409] [GatlingSystem-akka.actor.default-dispatcher-19] [akka://GatlingSystem/user/$d] Too many connections 10001
java.io.IOException: Too many connections 10001
demo.tgz (3.99 KB)
We have started cleaning up the gatling.conf file.
Many options could be both set up in protocol and in this file, too many ways to do the same thing, bringing confusion.
This has to be set up on the protocol now: https://github.com/excilys/gatling/blob/master/src/sphinx/http/http_protocol.rst#connections-sharing
I think we have to find a way to warn users of removed properties, let me think about it.
One option could be requiring a version to be set in each config file.
If the version number is not present or wrong, end the simulation, and print a message redirecting someone to migration notes.
Having an easy way of validating your config would be awesome, too. Or even migrating your config to a new version.
We’ll see about that in 2M5, as we intend to split the configuration into multiple files, so that Gatling gets more modular and plugins can use the same configuration facilities.
Just 'cause I thought it’d be fun to mention - I’ve got gatling generating 72K requests per second. That’s about 4.3 Mill requests / minute.
Could you share some details, please?
How many Gatling instances?
Do you share connections?
How many concurrent connections?
Mean response time?
Mean response payload?
What does the system under test look like?
One last question: did you notice some performance improvements in Gatling 2 snapshots in the last months?