
On Thu, Jan 9, 2014 at 5:33 PM, Dan Kenigsberg wrote:
The question of how much logging we should keep is a tough one. I, as a developer, would like to have as much as possible. For long-running busy systems, it has happened to me that the core bug was spotted in vdsm.log.67 or so.
However, I understand that verbosity has its price. To understand whether we are stable enough to change the defaults, I need volunteers: people who are willing to change their log level to INFO or WARNING, and see if they miss useful information from their logs.
When you make you log level higher, you can lower the number of kept log files, as they would not be filled as quick.
Would you, users@, help me with hard data?
Dan.
I can't tell about missing information because at the moment I didn't had big problmes since in my case I put logging to INFO where there was DEBUG on 17 December for both vdsm and supervdsm. This is a very calm infra composed by two nodes f19 oVirt 3.3.2 stable repo and GlusterFS datcenter. Only 3 VMS running about everytime Rotation has not yet overwritten old log files Under /var/log/vdsm now # du -sh . 137M . # ll supervdsm.log*|wc -l 106 # ll vdsm.log*|wc -l 113 For vdsm logs: before I had 5 daily generated files for an overall of about 100Mb uncompressed, now one file per day of about 14Mb uncompressed A newer log -rw-r--r--. 1 vdsm kvm 429380 Jan 7 12:00 vdsm.log.1.xz A before change log sequence -rw-r--r--. 1 vdsm kvm 660280 Dec 8 22:01 vdsm.log.36.xz -rw-r--r--. 1 vdsm kvm 659672 Dec 8 17:01 vdsm.log.37.xz -rw-r--r--. 1 vdsm kvm 662584 Dec 8 12:01 vdsm.log.38.xz -rw-r--r--. 1 vdsm kvm 655232 Dec 8 07:01 vdsm.log.39.xz -rw-r--r--. 1 vdsm kvm 657832 Dec 8 02:00 vdsm.log.40.xz For supervdsm logs no rotation yet after change, because size is set to 15M in logrotate conf files: Before change I had one file a day (uncompressed daily size was about 17-20Mb) # ls -lt super* -rw-r--r--. 1 root root 12620463 Jan 7 12:57 supervdsm.log -rw-r--r--. 1 root root 342736 Dec 16 16:01 supervdsm.log.1.xz -rw-r--r--. 1 root root 328952 Dec 15 14:00 supervdsm.log.2.xz -rw-r--r--. 1 root root 343360 Dec 14 13:01 supervdsm.log.3.xz -rw-r--r--. 1 root root 339244 Dec 13 11:00 supervdsm.log.4.xz -rw-r--r--. 1 root root 349012 Dec 12 09:00 supervdsm.log.5.xz I got a problem with SPM on 7 anuary and I correctly found in vdsm.log Thread-7000::INFO::2014-01-07 15:42:52,632::logUtils::44::dispatcher::(wrapper) Run and protect: disconnectStorageServer(domType=7, sp UUID='eb679feb-4da2-4fd0-a185-abbe459ffa70', conList=[{'port': '', 'connection': 'f18ovn01.mydomain:gvdata', 'iqn': '', 'portal' : '', 'user': '', 'vfs_type': 'glusterfs', 'password': '******', 'id': '9d01b8fa-853e-4720-990b-f86bdb7cfcbb'}], options=None) Thread-7001::INFO::2014-01-07 15:42:54,776::logUtils::44::dispatcher::(wrapper) Run and protect: getAllTasksStatuses(spUUID=None, opti ons=None) Thread-7001::ERROR::2014-01-07 15:42:54,777::task::850::TaskManager.Task::(_setError) Task=`2e948a29-fdaa-4049-ada4-421c6407b037`::Une xpected error Traceback (most recent call last): File "/usr/share/vdsm/storage/task.py", line 857, in _run return fn(*args, **kargs) File "/usr/share/vdsm/logUtils.py", line 45, in wrapper res = f(*args, **kwargs) File "/usr/share/vdsm/storage/hsm.py", line 2109, in getAllTasksStatuses raise se.SpmStatusError() SpmStatusError: Not SPM: () Thread-7001::INFO::2014-01-07 15:42:54,777::task::1151::TaskManager.Task::(prepare) Task=`2e948a29-fdaa-4049-ada4-421c6407b037`::abort ing: Task is aborted: 'Not SPM' - code 654 Thread-7001::ERROR::2014-01-07 15:42:54,778::dispatcher::67::Storage.Dispatcher.Protect::(run) {'status': {'message': 'Not SPM: ()', ' code': 654}} So I think it is good for me, otherwise I would have great difficulties identifying problems inside the big files produced Gianluca