Loghub-2.0 is an improved collection of large-scale annotated datasets for log parsing based on Loghub.
Based on Loghub-2.0, we propose a more comprehensive benchmark of log parsers. The detailed evaluation results could be found at RQ_experiments π.
| Software systems | # Annotated Logs (Loghub-2.0) | # Templates (Loghub-2.0) | # Templates (Loghub-2k) |
|---|---|---|---|
| Distributed systems | |||
| Hadoop | 179,993 | 236 | 114 |
| HDFS | 11,167,740 | 46 | 14 |
| OpenStack | 207,632 | 48 | 43 |
| Spark | 16,075,117 | 236 | 36 |
| Zookeeper | 74,273 | 89 | 50 |
| Supercomputer systems | |||
| BGL | 4,631,261 | 320 | 120 |
| HPC | 429,987 | 74 | 46 |
| Thunderbird | 16,601,745 | 1,241 | 149 |
| Operating systems | |||
| Linux | 23,921 | 338 | 118 |
| Mac | 100,314 | 626 | 341 |
| Server application | |||
| Apache | 51,977 | 29 | 6 |
| OpenSSH | 638,946 | 38 | 27 |
| Standalone software | |||
| HealthApp | 212,394 | 156 | 75 |
| Proxifier | 21,320 | 11 | 8 |
| Average | 3,601,187 | 249.1 | 81.9 |
Please first download the full datasets of Loghub-2.0 via Zenodo.
Then, you need to put these datasets into full_dataset/ following the format of 2k_dataset.
βββ 2k_dataset/ # the original Loghub-2k datasets
βββ full_dataset/ # unzip the Loghub-2.0 into this directory
β βββ post_process.py # we provide the heuristic roles used in our annotation of templates
βββ benchmark/
β βββ evaluation/
β βββ logparser/
β βββ old_benchmark/
β βββ LogPPT/ # contains the modified source code of LogPPT
β βββ UniParser/ # contains the source code of implemented UniParser
β βββ run_statistic_2k.sh # the script to run all statistic-based log parsers on Loghub-2k datasets
β βββ run_statistic_full.sh # the script to run all statistic-based log parsers on Loghub-2.0 datasets
βββ result/
β βββ ...... #
β βββ ...... # contains the output evaluation metric files and all parsed results
βββ RQ_experiments/ # contains the experimental results of RQs
β βββ RQ1/
β βββ RQ2/
β βββ RQ3/
βββ requirements.txt
βββ README.MD
Owing to the large scale of the benchmark in the experiments, the requirements of the benchmark of all log parsers are:
- At least 16GB memory.
- At least 100GB storage.
- GPU (for LogPPT and UniParser).
Installation
- Install
python >= 3.8 pip install -r requirements.txt
Running the entire benchmark using Loghub-2.0 datasets requires more than 48 hours to complete.
Note that if you would like to evaluate your parser, one can easily put their parsed results following the format as the files shown in result/, and run our evluation code.
If you woud like to re-run all parsers using Loghub-2.0, please follow our large-scale benchmarking steps.
We give a demo script to run Drain on both Loghub-2k and Loghub-2.0, this will takes about 2-3 hours.
cd benchmark/
./demo.shOne can follow the steps to evaluate all parsers using Loghub-2k or the proposed Loghub-2.0 datasets. The overall time cost is more than 48 hours.
- Run all statistic-based log parsers on Loghub-2k
cd benchmark/
./run_statistic_2k.sh- Run all statistic-based log parsers on Loghub-2.0
cd benchmark/
./run_statistic_full.sh-
Run Semantic-based log parsers: LogPPT & UniParser
Since these methods are quite different with other log parsers, and they requires a GPU to support efficient parsing, we seperate their environments from other log parsers. Please refer to the README file of LogPPT or UniParser to use one-click script to parse and evaluate each log parsers respectively.
If you use our labeled datasets for public research, please cite the following two papers:
- Loghub: Jieming Zhu, Shilin He, Pinjia He, Jinyang Liu, Michael R. Lyu. Loghub: A Large Collection of System Log Datasets for AI-driven Log Analytics. ISSRE, 2023.
- Loghub-2.0: Zhihan Jiang, Jinyang Liu, Junjie Huang, Yichen Li, Yintong Huo, Jiazhen Gu, Zhuangbin Chen, Jieming Zhu, Michael R. Lyu. A Large-scale Evaluation for Log Parsing Techniques: How Far are We? ISSTA, 2024.
In addition, if you use the souce code of our work for public research, please kindly cite the following paper for their efforts on logparser implementation:
- Jieming Zhu, Shilin He, Jinyang Liu, Pinjia He, Qi Xie, Zibin Zheng, Michael R. Lyu. Tools and Benchmarks for Automated Log Parsing. ICSE, 2019.