Glimpse on Tomcat performance tuning.

Production environment ;-)

Have You ever wondered about Tomcat configuration in production environment, or just let “this things” to the admins, or even worse, don’t care at all about it? If the answer is “Tomcat configuration ? I/We/Our client just installs tomcat and deploy our application. Why border about any additional configuration ?” You should read this post.

I will not write about all Tomcat’s configuration. It’s pointless. I just want to show some problems with performance with default Tomcat’s configuration in production enviroment. Especially if You are using Tomcat in as web server in internet, with many simultaneous clients and connections. In such cases performance and high responsivity is important.

1. Let’s start from logs. Standard Tomcat’s logs are configured to appear in two places: file and console. In production it’s pointless to have duplicate logs so first thing to gain some speed boost is to replace following line from logging.properties:

.handlers = 1catalina.org.apache.juli.FileHandler, java.util.logging.ConsoleHandler with this one: .handlers = 1catalina.org.apache.juli.FileHandler

2. Second thing to do with logs is to set max file size and protection from overflow. It’s also very easy. Just add new handler like following one:

catalina.java.util.logging.FileHandler

and configure it like this (max 4 filesx10Mb):

1catalina.java.util.logging.FileHandler.pattern = ${catalina.base}/logs/catalina.%g.log 1catalina.java.util.logging.FileHandler.limit = 10000000 1catalina.java.util.logging.FileHandler.count = 4

3. Last thing You have TO HAVE in production environment are asynchronous logs. Synchronous logging is far more time consuming then asynchronous one. Especially when You have numerous clients. Check if Your Tomcat is configured in proper way (I won’t write about this. Just search in web about log4j configuration. It’s lot of this there.)

4. That’s all about logging. Now something much more influent on connection speed-connectors. They are configured in server.xml under node.

Tomcat have 3 main connectors:

BIO – Blocking Java connector which is default one

APR – Uses native C code fo IO (very fast)

NIO – Non blocking connectror in Java (also faster than default)

The first BIO connector (“org.apache.coyote.http11.Http11Protocol”) is set as default one. Why ? Becouse in many cases such configuration it’s enough. Tomcat usually is used in intranets where it’s not required to handle high traffic volume. Moreover BIO connector is very stable.

But if our applications have to serve many http requests the blocking connector isn’t the best choice. So here comes ARP and NIO connector.

The first one (org.apache.coyote.http11.Http11AprProtocol) requires to compile native library (just search in google for ARP) and could be less stable than BIO connector. In exchange ARP connector is very fast, could handle requests simultanously in non blocking mode, have pooling of unlimited size and could handle unlimited threads (in theory, becouse threads are limited with CPU power)

Last connector – NIO (org.apache.coyote.http11.Http11NioProtocol) is something between ARP and BIO. It’s good choice if You don’t want to compile native libraries. NIO connector is also non blocking, little slower in reading static content than ARP, but far more configurable (pool size, no of threads etc).

5. Ok, so now We know, which connector should we choose, but every connector have to be set up in proper way. There are several parameters but the important ones are:

– maxThreads – typical from 150-800 (For BIO this is max nr of open connections)

– maxKeepAliveRequests – typical 1 or 100-250. For BIO this should be set to 1 to disable keep alive (only if we have high concurency and not using SSL). BIO connector automatically disables keep alive for high connection traffic

– connectionTimeout – typical 2000-60000 WARNING: default Tomcat has it set to 20 000! It’s to high for production environment. Good choice is to decrese it to 3000-5000 unless Your production env is working with slow clients. This parameters describes max time between TCP packets during blocking read/write

6. This is “almost” the end of tunning Tomcat for production. The last thing is to configure cache. Default cache is configured to 10 MB. You can set this a little more if You have a lot of static content. Also cache revalidation (standard 5 sec) should be tuned. How ? It’s difficult to say. The best way is to tune this parameters by own during tests.

That’s all. I hope I realized to everyone why not rely on standard Tomcat configuration.

You May Also Like

Log4j and MDC in Grails

Log4j provides very useful feature: MDC - mapped diagnostic context. It can be used to store data in context of current thread. It may sound scary a bit but idea is simple.

My post is based on post http://burtbeckwith.com/blog/?p=521 from Burt Beckwith's excellent blog, it's definitely worth checking if you are interested in Grails.

Short background story...


Suppose we want to do logging our brand new shopping system and we want to have in each log customer's shopping basket number. And our system can be used at once by many users who can perform many transactions, actions like adding items and so on. How can we achieve that? Of course we can add basket number in every place where we do some logging but this task would be boring and error-prone. 

Instead of this we can use MDC to store variable with basket number in map. 

In fact MDC can be treated as map of custom values for current thread that can be used by logger. 


How to do that with Grails?


Using MDC with Grails is quite simple. All we need to do is to create our own custom filter which works for given urls and puts our data in MDC.

Filters in Grails are classes in directory grails-app/conf/* which names end with *Filters.groovy postfix. We can create this class manually or use Grails command: 
grails create-filters info.rnowak.App.Basket

In result class named BasketFilters will be created in grails-app/conf/info/rnowak/UberApp.

Initially filter class looks a little bit empty:
class BasketFilters {
def filters = {
all(controller:'*', action:'*') {
before = {

}
after = { Map model ->

}
afterView = { Exception e ->

}
}
}
}
All we need to do is fill empty closures, modify filter properties and put some data into MDC.

all is the general name of our filter, as class BasketFilters (plural!) can contain many various filters. You can name it whatever you want, for this post let assume it will be named basketFilter

Another thing is change of filter parameters. According to official documentation (link) we can customize our filter in many ways. You can specify controller to be filtered, its actions, filtered urls and so on. In our example you can stay with default option where filter is applied to every action of every controller. If you are interested in filtering only some urls, use uri parameter with expression describing desired urls to be filtered.

Three closures that are already defined in template have their function and they are started in these conditions:

  • before - as name says, it is executed before filtered action takes place
  • after - similarly, it is called after the action
  • afterView - called after rendering of the actions view
Ok, so now we know what are these mysterious methods and when they are called. But what can be done within them? In official Grails docs (link again) under section 7.6.3 there is a list of properties that are available to use in filter.

With that knowledge, we can proceed to implementing filter.

Putting something into MDC in filter


What we want to do is quite easy: we want to retrieve basket number from parameters and put it into MDC in our filter:
class BasketFilters {
def filters = {
basketFilter(controller:'*', action:'*') {
before = {
MDC.put("basketNumber", params.basketNumber ?: "")
}
after = { Map model ->
MDC.remove("basketNumber")
}
}
}
}

We retrieve basket number from Grails params map and then we put in map under specified key ("basketNumber" in this case), which will be later used in logger conversion pattern. It is important to remove custom value after processing of action to avoid leaks.

So we are putting something into MDC. But how make use of it in logs?


We can refer to custom data in MDC in conversion patter using syntax: %X{key}, where key is our key we used in filter to put data, like:
def conversionPattern = "%d{yyyy-MM-dd HH:mm:ss} %-5p %t [%c{1}] %X{basketNumber} - %m%n"


And that's it :) We've put custom data in log4j MDC and successfully used it in logs to display interesting values.