Regenerate DWH ovirt_engine_history

Hello to all and happy new year. My question is "simple". I need to "reset" the ovirt_engine_history database. I tried to use: engine-setup --reconfigure-optional-components after removing: - ovirt_engine_history - set to "False" OVESETUP_DWH_CORE/enable=bool:True OVESETUP_GRAFANA_CORE/enable=bool:True in /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf (obviously all in "Global Maintenance") currently I have a system where I have an initialized ovirt_engine_history but no engine is writing in it. I tried to follow (in some manner): https://www.ovirt.org/documentation/data_warehouse_guide/ In the ovirt engine's log there is no exception... so I cannot address the issue. I need help to reinitialize the db... I don't need the history.. only from now on. Thank you for your help but no one is

Sorry, some environment: - ovirt hosted-engine (self hosted) - [root@ovirt-engine ~]# rpm -qa | grep engine ovirt-engine-setup-plugin-ovirt-engine-4.5.4-1.el8.noarch ovirt-engine-extension-aaa-ldap-1.4.6-1.el8.noarch ovirt-engine-backend-4.5.4-1.el8.noarch python3-ovirt-engine-sdk4-4.6.0-1.el8.x86_64 ovirt-engine-dwh-4.5.7-1.el8.noarch python38-ovirt-engine-sdk4-4.5.1-1.el8.x86_64 ovirt-engine-setup-plugin-ovirt-engine-common-4.5.4-1.el8.noarch ovirt-engine-websocket-proxy-4.5.4-1.el8.noarch ovirt-engine-4.5.4-1.el8.noarch ovirt-engine-dwh-grafana-integration-setup-4.5.7-1.el8.noarch ovirt-engine-wildfly-overlay-24.0.1-1.el8.noarch ovirt-engine-setup-plugin-cinderlib-4.5.4-1.el8.noarch ovirt-engine-setup-4.5.4-1.el8.noarch ovirt-engine-vmconsole-proxy-helper-4.5.4-1.el8.noarch ovirt-engine-restapi-4.5.4-1.el8.noarch ovirt-engine-extension-aaa-misc-1.1.1-1.el8.noarch ovirt-engine-extensions-api-1.0.1-1.el8.noarch ovirt-engine-ui-extensions-1.3.7-1.el8.noarch ovirt-engine-dwh-setup-4.5.7-1.el8.noarch ovirt-engine-extension-aaa-jdbc-1.3.0-1.el8.noarch ovirt-engine-wildfly-24.0.1-1.el8.x86_64 ovirt-engine-metrics-1.6.1-1.el8.noarch python3-ovirt-engine-lib-4.5.4-1.el8.noarch ovirt-engine-setup-plugin-imageio-4.5.4-1.el8.noarch ovirt-engine-setup-plugin-websocket-proxy-4.5.4-1.el8.noarch ovirt-engine-tools-backup-4.5.4-1.el8.noarch ovirt-engine-tools-4.5.4-1.el8.noarch ovirt-engine-keycloak-15.0.2-6.el8.noarch ovirt-engine-keycloak-setup-15.0.2-6.el8.noarch ovirt-engine-setup-base-4.5.4-1.el8.noarch ovirt-engine-webadmin-portal-4.5.4-1.el8.noarch ovirt-engine-setup-plugin-vmconsole-proxy-helper-4.5.4-1.el8.noarch ovirt-engine-extension-aaa-ldap-setup-1.4.6-1.el8.noarch python39-ovirt-engine-sdk4-4.6.0-1.el8.x86_64 ovirt-engine-dbscripts-4.5.4-1.el8.noarch

Hi, On Sun, Jan 1, 2023 at 10:19 PM Diego Ercolani <diego.ercolani@ssis.sm> wrote:
Hello to all and happy new year. My question is "simple". I need to "reset" the ovirt_engine_history database. I tried to use: engine-setup --reconfigure-optional-components after removing: - ovirt_engine_history - set to "False" OVESETUP_DWH_CORE/enable=bool:True OVESETUP_GRAFANA_CORE/enable=bool:True in /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf
(obviously all in "Global Maintenance") currently I have a system where I have an initialized ovirt_engine_history but no engine is writing in it.
It's not the engine that is writing there, it's dwhd. The engine only reads. Did you check /var/log/ovirt-engine-dwh/ ?
I tried to follow (in some manner): https://www.ovirt.org/documentation/data_warehouse_guide/
In the ovirt engine's log there is no exception... so I cannot address the issue.
I need help to reinitialize the db... I don't need the history.. only from now on.
What you did might be enough, in principle, but we do not have it documented/tested, IIRC. How did you remove ovirt_engine_history? Did you remove both db and user? And lines from pghba.conf? You might need to remove also /etc/ovirt-engine/engine.conf.d/10-setup-dwh-database.conf , before running engine-setup. That's the file the engine uses to get credentials for the dwh db. Good luck and best regards, -- Didi

Thank you for your infos. > It's not the engine that is writing there, it's dwhd. The engine only > reads. Did you check /var/log/ovirt-engine-dwh/ ? What is confusing me are these line in /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log lastErrorSent|2011-07-03 12:46:47.000000 etlVersion|4.5.7 dwhAggregationDebug|false dwhUuid|53878f7e-cb3d-45cc-9edf-9e389414ce6e ovirtEngineHistoryDbDriverClass|org.postgresql.Driver ovirtEngineHistoryDbPassword|********************** 2022-12-30 21:45:01|6kFI8I|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 2022-12-30 22:00:01|HuESUa|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 2022-12-30 22:15:01|i99aVI|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 > > > What you did might be enough, in principle, but we do not have it > documented/tested, IIRC. It should be a "nice to have feature".... > > How did you remove ovirt_engine_history? Did you remove both db and > user? And lines from pghba.conf? > > You might need to remove also > /etc/ovirt-engine/engine.conf.d/10-setup-dwh-database.conf , before > running engine-setup. That's the file the engine uses to get > credentials for the dwh db. I noticed that engine-setup wrote the correct credentials even if you don't remove the file > > Good luck and best regards, I decided to restart the clean process using your info so, I'm going to trace what I do during the process in the following lines: 1. put the cluster in global maintenance 2. systemctl stop ovirt-engine-dwhd, systemctl stop ovirt-engine, systemctl stop grafana-server 3. verify what process i using the dwh db: SELECT * FROM pg_stat_activity where datname='ovirt_engine_history'; 4. "Clean" the database: DROP DATABASE ovirt_engine_history; DROP ROLE ovirt_engine_history; DROP ROLE ovirt_engine_history_grafana; 5. "Clean" the setup environment: --- /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf.bak 2022-12-30 12:52:19.810983574 +0000 +++ /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf 2023-01-02 09:48:43.642373549 +0000 @@ -19,10 +19,10 @@ OVESETUP_CONFIG/adminUserId=str:9c1af7f1-5795-432a-b971-9883e0bed76d OVESETUP_OVN/ovirtProviderOvn=bool:True OVESETUP_OVN/ovirtProviderOvnId=str:e6b92384-b112-40e0-8d6f-2c6e4536cd1a -OVESETUP_DWH_CORE/enable=bool:True +OVESETUP_DWH_CORE/enable=bool:False OVESETUP_DWH_CONFIG/remoteEngineConfigured=bool:False OVESETUP_DWH_CONFIG/scale=str:2 -OVESETUP_GRAFANA_CORE/enable=bool:True +OVESETUP_GRAFANA_CORE/enable=bool:False OVESETUP_GRAFANA_CORE/grafanaDbCreatedByUs=none:None OVESETUP_KEYCLOAK_CONFIG/ovirtAdminUser=str:admin@ovirt OVESETUP_VMCONSOLE_PROXY_CONFIG/vmconsoleProxyConfig=bool:True rm /etc/ovirt-engine/engine.conf.d/10-setup-dwh-database.conf --- /var/lib/pgsql/data/pg_hba.conf.bak 2023-01-02 09:52:34.153904851 +0000 +++ /var/lib/pgsql/data/pg_hba.conf 2023-01-02 09:53:08.784158670 +0000 @@ -82,14 +82,6 @@ # "local" is for Unix domain socket connections only local all all peer -host ovirt_engine_history ovirt_engine_history_grafana 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history_grafana ::0/0 md5 -host ovirt_engine_history ovirt_engine_history 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history ::0/0 md5 host engine engine 0.0.0.0/0 md5 host engine engine ::0/0 md5 # IPv4 local connections: @@ -99,18 +91,9 @@ # Allow replication connections from localhost, by a user with the # replication privilege. local replication all peer -host ovirt_engine_history ovirt_engine_history_grafana 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history_grafana ::0/0 md5 -host ovirt_engine_history ovirt_engine_history 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history ::0/0 md5 host engine engine 0.0.0.0/0 md5 host engine engine ::0/0 md5 host replication all 127.0.0.1/32 ident host replication all ::1/128 ident host all diego 192.168.9.0/24 password host all axel 192.168.9.0/24 trust 6. launch engine-setup --reconfigure-optional-components: --== CONFIGURATION PREVIEW ==-- Default SAN wipe after delete : False Host FQDN : ovirt-engine.ovirt Firewall manager : firewalld Update Firewall : True Set up Cinderlib integration : False Engine database host : localhost Engine database port : 5432 Engine database secured connection : False Engine database host name validation : False Engine database name : engine Engine database user name : engine Engine installation : True PKI organization : ovirt Set up ovirt-provider-ovn : True DWH installation : True DWH database host : localhost DWH database port : 5432 DWH database secured connection : False DWH database host name validation : False DWH database name : ovirt_engine_history Configure local DWH database : True Grafana integration : True Grafana database user name : ovirt_engine_history_grafana Configure VMConsole Proxy : True Configure WebSocket Proxy : True The engine-setup seem to have a successful run.... the ovirt_engine_history seem to work, no error in logs: [root@ovirt-engine ~]# tail -10 /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log deleteIncrement|10 timeBetweenErrorEvents|300000 hoursToKeepSamples|24 deleteMultiplier|1000 lastErrorSent|2011-07-03 12:46:47.000000 etlVersion|4.5.7 dwhAggregationDebug|false dwhUuid|53878f7e-cb3d-45cc-9edf-9e389414ce6e ovirtEngineHistoryDbDriverClass|org.postgresql.Driver ovirtEngineHistoryDbPassword|********************** [root@ovirt-engine ~]# grep -e "DWH_DB_\(USER\|PASSWORD\)" /etc/ovirt-engine-dwh/ovirt-engine-dwhd.conf.d/10-setup-database.conf DWH_DB_USER="ovirt_engine_history" DWH_DB_PASSWORD="7l0W2uqqoy3EYS1gD0VRwE" [root@ovirt-engine ~]# psql -U ovirt_engine_history -h ovirt-engine.ovirt -W ovirt_engine_history Password: psql (12.12) Type "help" for help. ovirt_engine_history-> \du List of roles Role name | Attributes | Member of ------------------------------+------------------------------------------------------------+----------- axel | Superuser, Create role, Create DB | {} diego | Superuser, Create role, Create DB | {} engine | Superuser | {} ovirt_engine_history | | {} ovirt_engine_history_grafana | | {} postgres | Superuser, Create role, Create DB, Replication, Bypass RLS | {} I'm currently waiting...

On Mon, Jan 2, 2023 at 12:09 PM Diego Ercolani <diego.ercolani@ssis.sm> wrote:
Thank you for your infos.
It's not the engine that is writing there, it's dwhd. The engine only reads. Did you check /var/log/ovirt-engine-dwh/ ? What is confusing me are these line in /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log lastErrorSent|2011-07-03 12:46:47.000000 etlVersion|4.5.7 dwhAggregationDebug|false dwhUuid|53878f7e-cb3d-45cc-9edf-9e389414ce6e ovirtEngineHistoryDbDriverClass|org.postgresql.Driver ovirtEngineHistoryDbPassword|********************** 2022-12-30 21:45:01|6kFI8I|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 2022-12-30 22:00:01|HuESUa|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 2022-12-30 22:15:01|i99aVI|Yi5fBA|8mXcq0|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704
The process is like this: 1. The engine routinely updates its own database 2. dwhd reads the engine DB. If it notices updates, it updates its own DB. If it does not, after some time it emits the above lines to the log. Perhaps the problem is that dwhd fails to look at the engine DB? Or uses wrong creds? They are in /etc/ovirt-engine-dwh/ovirt-engine-dwhd.conf.d/10-setup-database.conf , keys starting with ENGINE_DB_ (those starting with DWH_DB_ are for its own db).
What you did might be enough, in principle, but we do not have it documented/tested, IIRC.
It should be a "nice to have feature"....
We had it open for many years but never finished... https://bugzilla.redhat.com/show_bug.cgi?id=1060529 https://bugzilla.redhat.com/show_bug.cgi?id=1781095
How did you remove ovirt_engine_history? Did you remove both db and user? And lines from pghba.conf?
You might need to remove also /etc/ovirt-engine/engine.conf.d/10-setup-dwh-database.conf , before running engine-setup. That's the file the engine uses to get credentials for the dwh db.
I noticed that engine-setup wrote the correct credentials even if you don't remove the file
Good luck and best regards,
I decided to restart the clean process using your info so, I'm going to trace what I do during the process in the following lines: 1. put the cluster in global maintenance 2. systemctl stop ovirt-engine-dwhd, systemctl stop ovirt-engine, systemctl stop grafana-server 3. verify what process i using the dwh db: SELECT * FROM pg_stat_activity where datname='ovirt_engine_history'; 4. "Clean" the database: DROP DATABASE ovirt_engine_history; DROP ROLE ovirt_engine_history; DROP ROLE ovirt_engine_history_grafana; 5. "Clean" the setup environment:
--- /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf.bak 2022-12-30 12:52:19.810983574 +0000 +++ /etc/ovirt-engine-setup.conf.d/20-setup-ovirt-post.conf 2023-01-02 09:48:43.642373549 +0000 @@ -19,10 +19,10 @@ OVESETUP_CONFIG/adminUserId=str:9c1af7f1-5795-432a-b971-9883e0bed76d OVESETUP_OVN/ovirtProviderOvn=bool:True OVESETUP_OVN/ovirtProviderOvnId=str:e6b92384-b112-40e0-8d6f-2c6e4536cd1a -OVESETUP_DWH_CORE/enable=bool:True +OVESETUP_DWH_CORE/enable=bool:False OVESETUP_DWH_CONFIG/remoteEngineConfigured=bool:False OVESETUP_DWH_CONFIG/scale=str:2 -OVESETUP_GRAFANA_CORE/enable=bool:True +OVESETUP_GRAFANA_CORE/enable=bool:False OVESETUP_GRAFANA_CORE/grafanaDbCreatedByUs=none:None OVESETUP_KEYCLOAK_CONFIG/ovirtAdminUser=str:admin@ovirt OVESETUP_VMCONSOLE_PROXY_CONFIG/vmconsoleProxyConfig=bool:True
rm /etc/ovirt-engine/engine.conf.d/10-setup-dwh-database.conf
--- /var/lib/pgsql/data/pg_hba.conf.bak 2023-01-02 09:52:34.153904851 +0000 +++ /var/lib/pgsql/data/pg_hba.conf 2023-01-02 09:53:08.784158670 +0000 @@ -82,14 +82,6 @@
# "local" is for Unix domain socket connections only local all all peer -host ovirt_engine_history ovirt_engine_history_grafana 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history_grafana ::0/0 md5 -host ovirt_engine_history ovirt_engine_history 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history ::0/0 md5 host engine engine 0.0.0.0/0 md5 host engine engine ::0/0 md5 # IPv4 local connections: @@ -99,18 +91,9 @@ # Allow replication connections from localhost, by a user with the # replication privilege. local replication all peer -host ovirt_engine_history ovirt_engine_history_grafana 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history_grafana ::0/0 md5 -host ovirt_engine_history ovirt_engine_history 0.0.0.0/0 md5 -host ovirt_engine_history ovirt_engine_history ::0/0 md5 host engine engine 0.0.0.0/0 md5 host engine engine ::0/0 md5 host replication all 127.0.0.1/32 ident host replication all ::1/128 ident host all diego 192.168.9.0/24 password host all axel 192.168.9.0/24 trust
6. launch engine-setup --reconfigure-optional-components: --== CONFIGURATION PREVIEW ==--
Default SAN wipe after delete : False Host FQDN : ovirt-engine.ovirt Firewall manager : firewalld Update Firewall : True Set up Cinderlib integration : False Engine database host : localhost Engine database port : 5432 Engine database secured connection : False Engine database host name validation : False Engine database name : engine Engine database user name : engine Engine installation : True PKI organization : ovirt Set up ovirt-provider-ovn : True DWH installation : True DWH database host : localhost DWH database port : 5432 DWH database secured connection : False DWH database host name validation : False DWH database name : ovirt_engine_history Configure local DWH database : True Grafana integration : True Grafana database user name : ovirt_engine_history_grafana Configure VMConsole Proxy : True Configure WebSocket Proxy : True
The engine-setup seem to have a successful run.... the ovirt_engine_history seem to work, no error in logs: [root@ovirt-engine ~]# tail -10 /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log deleteIncrement|10 timeBetweenErrorEvents|300000 hoursToKeepSamples|24 deleteMultiplier|1000 lastErrorSent|2011-07-03 12:46:47.000000 etlVersion|4.5.7 dwhAggregationDebug|false dwhUuid|53878f7e-cb3d-45cc-9edf-9e389414ce6e ovirtEngineHistoryDbDriverClass|org.postgresql.Driver ovirtEngineHistoryDbPassword|**********************
[root@ovirt-engine ~]# grep -e "DWH_DB_\(USER\|PASSWORD\)" /etc/ovirt-engine-dwh/ovirt-engine-dwhd.conf.d/10-setup-database.conf DWH_DB_USER="ovirt_engine_history" DWH_DB_PASSWORD="7l0W2uqqoy3EYS1gD0VRwE"
(Better not expose passwords to the Internet)
[root@ovirt-engine ~]# psql -U ovirt_engine_history -h ovirt-engine.ovirt -W ovirt_engine_history Password: psql (12.12) Type "help" for help. ovirt_engine_history-> \du List of roles Role name | Attributes | Member of ------------------------------+------------------------------------------------------------+----------- axel | Superuser, Create role, Create DB | {} diego | Superuser, Create role, Create DB | {} engine | Superuser | {} ovirt_engine_history | | {} ovirt_engine_history_grafana | | {} postgres | Superuser, Create role, Create DB, Replication, Bypass RLS | {}
I'm currently waiting...
Sorry. If all else fails, perhaps try this: 1. Setup a test engine+dwh, make sure it works (and you do not see log lines about engine not updating stats). 2. Study this setup - check what process connects to what DB using what creds, and where these are saved, and compare to your own setup. Best regards, -- Didi

All the files seem to be correctly intializated. The only doubt is in the last directory you addressed: /etc/ovirt-engine-dwh/ovirt-engine-dwhd.conf.d/ there is a file: [root@ovirt-engine ovirt-engine-dwhd.conf.d]# ls -ltr total 28 -rw-r--r--. 1 root root 223 Oct 5 09:17 README -rw-------. 1 ovirt ovirt 83 Dec 30 12:51 10-setup-scale.conf -rw-------. 1 ovirt ovirt 727 Dec 30 12:52 10-setup-database.conf.20230102095726 -rw-r-----. 1 root ovirt 415 Dec 30 12:52 10-setup-grafana-database.conf.20230102095726 -rw-r--r--. 1 root root 46 Dec 30 12:52 10-setup-uuid.conf -rw-------. 1 ovirt ovirt 727 Jan 2 09:57 10-setup-database.conf -rw-r-----. 1 root ovirt 415 Jan 2 09:57 10-setup-grafana-database.conf 10-setup-uuid.conf have the timestamp of dec 30... what is used for?

I found the reference on that file: https://github.com/oVirt/ovirt-dwh/blob/master/docs/Notes-about-single-dwhd It's only to notice that I veryfied the contents of dwh_history_timeskeeping table @engine db and the dwhUuid it's consistent with the one in the 10-setup-uuid.conf file While ovirt-engine-dwh is running the value DwhCurrentlyRunning is correctly set to "1" and when I issue systemctl stop ovirt-engine-dwhd the parameter is set to 0, so the dwhd correctly can talk with engine db... I suppose. The DisconnectDwh in the vdc_options table is correctly (?) set to its default value "0"

Finally it worked: After the step previous described: 1. put cluster in global maintenance 2. stop ovirt-engine and ovirt-engine-dwhd 3. in the table dwh_history_timekeeping @enginedb I changed the dwhUuid 4. launched engine-setup, the engine-setup asked to disconnect a "fantomatic" DWH (I answered "YES") 5. engine setup reconfigure things and set back the dwhUuid to the previous state 6. stopped ovirt-engine and ovirt-engine-dwhd because I saw that in the same table I had timestamp in the future (year 2177), so I removed the erroneous timestamp (putting to [NULL]) 7. in the /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log I saw: Setting a value for the key "lastErrorSent" has failed. Error message: null Setting a value for the key "lastErrorSent" has failed. Error message: null so I put in "consistent" value: This is my set: INSERT INTO public.dwh_history_timekeeping (var_name,var_value,var_datetime) VALUES ('dwhHostname','ovirt-engine.ovirt',NULL), ('dwhUuid','53878f7e-cb3d-45cc-9edf-9e389414ce6e',NULL), ('lastSampling',NULL,'2023-01-02 13:32:00.2+01'), ('lastSync',NULL,'2023-01-02 13:31:00+01'), ('lastErrorSent','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318+01'), ('lastOsinfoSync','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318+01'), ('lastFullHostCheck','2001-01-01 02:01:00.000 +0100','2023-01-02 13:03:00+01'), ('lastOsinfoUpdate','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318455+01'), ('heartBeat',NULL,'2023-01-02 13:31:59.082+01'), ('timesFailed','0',NULL); INSERT INTO public.dwh_history_timekeeping (var_name,var_value,var_datetime) VALUES ('DwhCurrentlyRunning','1',NULL);

On Mon, Jan 2, 2023 at 2:33 PM Diego Ercolani <diego.ercolani@ssis.sm> wrote:
Finally it worked: After the step previous described: 1. put cluster in global maintenance 2. stop ovirt-engine and ovirt-engine-dwhd 3. in the table dwh_history_timekeeping @enginedb I changed the dwhUuid 4. launched engine-setup, the engine-setup asked to disconnect a "fantomatic" DWH (I answered "YES") 5. engine setup reconfigure things and set back the dwhUuid to the previous state 6. stopped ovirt-engine and ovirt-engine-dwhd because I saw that in the same table I had timestamp in the future (year 2177),
This might hint at the issue. I do not know the relevant logic well, sorry. Perhaps you had some local time issue which caused this line to be written? Perhaps there is then logic to not update/ignore/whatever stuff that's written "in the past". Anyway, good catch!
so I removed the erroneous timestamp (putting to [NULL]) 7. in the /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log I saw: Setting a value for the key "lastErrorSent" has failed. Error message: null Setting a value for the key "lastErrorSent" has failed. Error message: null so I put in "consistent" value:
This is my set: INSERT INTO public.dwh_history_timekeeping (var_name,var_value,var_datetime) VALUES ('dwhHostname','ovirt-engine.ovirt',NULL), ('dwhUuid','53878f7e-cb3d-45cc-9edf-9e389414ce6e',NULL), ('lastSampling',NULL,'2023-01-02 13:32:00.2+01'), ('lastSync',NULL,'2023-01-02 13:31:00+01'), ('lastErrorSent','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318+01'), ('lastOsinfoSync','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318+01'), ('lastFullHostCheck','2001-01-01 02:01:00.000 +0100','2023-01-02 13:03:00+01'), ('lastOsinfoUpdate','2023-01-02 12:47:15.318 +0100','2023-01-02 12:47:15.318455+01'), ('heartBeat',NULL,'2023-01-02 13:31:59.082+01'), ('timesFailed','0',NULL); INSERT INTO public.dwh_history_timekeeping (var_name,var_value,var_datetime) VALUES ('DwhCurrentlyRunning','1',NULL);
Best regards, -- Didi

On Mon, Jan 2, 2023 at 1:38 PM Diego Ercolani <diego.ercolani@ssis.sm> wrote:
I found the reference on that file: https://github.com/oVirt/ovirt-dwh/blob/master/docs/Notes-about-single-dwhd
I didn't remember I wrote it :-(.
It's only to notice that I veryfied the contents of dwh_history_timeskeeping table @engine db and the dwhUuid it's consistent with the one in the 10-setup-uuid.conf file
While ovirt-engine-dwh is running the value DwhCurrentlyRunning is correctly set to "1" and when I issue systemctl stop ovirt-engine-dwhd the parameter is set to 0, so the dwhd correctly can talk with engine db... I suppose.
The DisconnectDwh in the vdc_options table is correctly (?) set to its default value "0"
All looks good to me. Are you there was a problem (I saw your other mail saying you fixed it)? You get the very same log lines about the engine not updating statistics even if there is no problem at all - this is generally what happens on a new engine with no entities being updated, left doing nothing. Best regards, -- Didi

Hello, I reinstalled the hosted-engine (see https://lists.ovirt.org/archives/list/users@ovirt.org/thread/QOJYO43SVOMCX6N...) I found a "glitch" in the grafana, somehow navigating the grafana environment (brand new) have put in inconsistent state the ovirt_history_db. The symptoms are: 1. a complain of the ovirt event view: 17 feb 2176, 08:47:17 ETL service sampling has encountered an error. Please consult the service log for more details. 17 feb 2176, 08:47:06 VDSM ovirt-node3.ovirt command Get Host Statistics failed: Connection timeout for host 'ovirt-node3.ovirt', last response arrived 1004236 ms ago. the dashboard was presenting no data so I checked the engine db: dwhHostname ovirt-engine.ovirt dwhUuid 93c4fc7f-74ef-43af-8529-3de33f2f763d lastOsinfoSync 2023-01-19 12:49:10.762 +0100 lastErrorSent 2176-02-17 08:47:17.126 +0100 DwhCurrentlyRunning 1 lastSync 2023-01-19 13:01:01.000 +0100 lastSampling 2177-03-03 22:11:13.023 +0100 lastOsinfoUpdate 2023-01-19 13:21:50.839 +0100 lastFullHostCheck 2023-01-19 12:56:01.000 +0100 timesFailed 129 heartBeat 2023-01-20 14:29:57.401 +0100 changing the errata entries (year>2176) I solved the problem: dwhHostname ovirt-engine.ovirt dwhUuid 93c4fc7f-74ef-43af-8529-3de33f2f763d NULL lastOsinfoSync NULL 2023-01-19 13:21:50.839 +0100 lastFullHostCheck NULL 2023-01-20 14:35:17.000 +0100 lastSync NUL 2023-01-20 14:57:17.000 +0100 DwhCurrentlyRunning 1 NULL lastOsinfoUpdate NULL 2023-01-19 13:21:50.839 +0100 timesFailed 0 NULL lastSampling 2023-01-20 14:58:17.300 +0100 heartBeat 2023-01-20 14:58:28.506 +0100 lastErrorSent NULL NULL what infos you need? Diego
participants (2)
-
Diego Ercolani
-
Yedidyah Bar David