Are there any existing batch log file aggregation solutions?
Asked Answered
P

4

7

I wish to export from multiple nodes log files (in my case apache access and error logs) and aggregate that data in batch, as a scheduled job. I have seen multiple solutions that work with streaming data (i.e think scribe). I would like a tool that gives me the flexibility to define the destination. This requirement comes from the fact that I want to use HDFS as the destination.

I have not been able to find a tool that supports this in batch. Before re-creating the wheel I wanted to ask the StackOverflow community for their input.

If a solution exists already in python that would be even better.

Portal answered 1/3, 2010 at 20:33 Comment(0)
P
1

we use http://mergelog.sourceforge.net/ to merge all our apache logs..

Papeterie answered 30/7, 2010 at 20:15 Comment(0)
G
0

take a look at Zomhg, its an aggregation/reporting system for log files using Hbase and Hdfs: http://github.com/zohmg/zohmg

Godolphin answered 31/3, 2010 at 20:21 Comment(0)
A
0

Scribe can meet your requirements, there's a version (link) of scribe that can aggregate logs from multiple sources, and after reaching given threshold it stores everything in HDFS. I've used it and it works very well. Compilation is quite complicated, so if you have any problems ask a question.

Anywhere answered 23/6, 2010 at 19:24 Comment(0)
S
-1

PiCloud may help.

The PiCloud Platform gives you the freedom to develop your algorithms and software without sinking time into all of the plumbing that comes with provisioning, managing, and maintaining servers.

Saltatory answered 6/3, 2010 at 2:12 Comment(0)

© 2022 - 2024 — McMap. All rights reserved.