13

Right now in my application,at certain points we are logging some heavy stuff into the log files.

Basically only for logging we are creating JSON of the data available and then logging into Log files.This is business requirement to log data in JSON format .

Now creating JSON from the data available and then logging to FILE takes lot of time and impacts the original request return time. Now idea is to improve the sitation .

One of the things that we have discussed is to create a thread pool using

Executors.newSingleThreadExecutor() 

in our code and then submitting the task to it which does the conversion of data into JSON and subsequent logging.

Is it a good approach to do this ?As we are managing the thread pool itself ,is it going to create some issues?

I would appreciate if someone can share better solutions. Someway to use Log4j for this .I tried to use AsyncAppender but didnt achieve any desired result. We are using EJB 3,Jboss 5.0,Log4j,java6.

Rips
  • 1,852
  • 1
  • 23
  • 41

5 Answers5

7

I believe you are on right track in terms of using a separate thread pool for logging. In lot of products you will see the asynchronous logging feature. Logs are accumulated and pushed to log files using a separate thread than the request thread. Especially in prodcution environments, where are millions of incoming request and your response time need to be less than few seconds. You cannot afford anything such as logging to slow down the system. So the approach used is to add logs in a memory buffer and push them asynchronously in reasonably sized chunks.

A word of caution while using thread pool for logging As multiple threads will be working on the log file(s) and on a memory log buffer, you need to be careful about the logging. You need to add logs in a FIFO kind of a buffer to be sure that logs are printed in the log files sorted by time stamp. Also make sure the file access is synchronized and you don't run into situation where log file is all upside down or messed up.

Juned Ahsan
  • 63,914
  • 9
  • 87
  • 123
  • Thanks for giving me confidence to go ahead in the right direction.But my only worry is I am having using application created threadpool...will that create any issue or is there any way I can leverage jboss thread pool for this. – Rips Jun 11 '13 at 06:38
6

Have a look at Logback,AsyncAppender it already provide separate threadpool, queue etc and is easily configurable, it almost do the same as you are doing, but saves you from re-inventing the wheel.

  • Is there any difference between this and log4j.AsyncAppender ?Also as part of this I dont only want to log asychronosuly but also convert data into JSON. – Rips Jun 10 '13 at 11:11
  • @Rips If you have other task also(like data conversion), then your solution is also ok. Regarding log4j vs logback, log4j development is stopped and logback is active. For more details about log4j vs logback have a look at this [link](http://stackoverflow.com/questions/178215/log4j-vs-logback) – Alankar Srivastava Jun 10 '13 at 11:22
3

Is using MongoDB for logging considered?

  1. MongoDB inserts can be done asynchronously. One wouldn’t want a user’s experience to grind to a halt if logging were slow, stalled or down. MongoDB provides the ability to fire off an insert into a log collection and not wait for a response code. (If one wants a response, one calls getLastError() — we would skip that here.)
  2. Old log data automatically LRU’s out. By using capped collections, we preallocate space for logs, and once it is full, the log wraps and reuses the space specified. No risk of filling up a disk with excessive log information, and no need to write log archival / deletion scripts.
  3. It’s fast enough for the problem. First, MongoDB is very fast in general, fast enough for problems like this. Second, when using a capped collection, insertion order is automatically preserved: we don’t need to create an index on timestamp. This makes things even faster, and is important given that the logging use case has a very high number of writes compared to reads (opposite of most database problems).
  4. Document-oriented / JSON is a great format for log information. Very flexible and “schemaless” in the sense we can throw in an extra field any time we want.
rajesh
  • 3,081
  • 4
  • 28
  • 52
1

There is also log4j 2: http://logging.apache.org/log4j/2.x/manual/async.html Additionally read this article about why it is so fast: http://www.grobmeier.de/log4j-2-performance-close-to-insane-20072013.html#.UzwywI9Bow4

Karussell
  • 16,303
  • 14
  • 88
  • 188
0

You can also try CoralLog to asynchronously log data using the disruptor pattern. That way you spend minimum time in the logger thread and all the hard work is passed to the thread doing the actual file I/O. It also provides Memory Mapped Files to speed up the consumer thread and reduce queue contention.

Disclaimer: I am one of the developers of CoralLog

rdalmeida
  • 1,680
  • 10
  • 15