On Tue, Jan 16, 2018 at 2:21 PM, Gianluca Cecchi <gianluca.cecchi@gmail.com> wrote:On Tue, Jan 16, 2018 at 6:48 AM, Fred Rolland <frolland@redhat.com> wrote:Hi,I will look into it.Is it also not working also for non-iso images?Thanks,FredHello,I get the same with a disk.I have tried with a raw disk of size 1Gb.When I upload I can set the size (why?) while with the iso image I could not.The disk is recognized as "data" in upload window (the iso fie was correctly recognized as "iso"), but in image-proxy.log I get(Thread-42 ) INFO 2018-01-16 14:15:50,530 web:95:web:(log_start) START [192.168.150.101] GET /info/(Thread-42 ) INFO 2018-01-16 14:15:50,532 web:102:web:(log_finish) FINISH [192.168.150.101] GET /info/: [200] 20 (0.00s)(Thread-43 ) INFO 2018-01-16 14:16:12,659 web:95:web:(log_start) START [192.168.150.105] PUT /tickets/(Thread-43 ) INFO 2018-01-16 14:16:12,661 auth2:170:auth2:(add_signed_ticket) Adding new ticket: <Ticket id=u'81569ab3-1b92- 4744-8a58-f1948afa20b7', url=u'https://ovirt01.localdomain.local:54322 ' at 0x7f048c03c610>(Thread-43 ) INFO 2018-01-16 14:16:12,662 web:102:web:(log_finish) FINISH [192.168.150.105] PUT /tickets/: [200] 0 (0.00s)(Thread-44 ) INFO 2018-01-16 14:16:13,800 web:95:web:(log_start) START [192.168.150.101] OPTIONS /images/81569ab3-1b92-4744-8a58-f1948afa20b7(Thread-44 ) INFO 2018-01-16 14:16:13,814 web:102:web:(log_finish) FINISH [192.168.150.101] OPTIONS /images/81569ab3-1b92-4744-8a58-f1948afa20b7: [204] 0 (0.02s)(Thread-45 ) INFO 2018-01-16 14:16:13,876 web:95:web:(log_start) START [192.168.150.101] PUT /images/81569ab3-1b92-4744-8a58 -f1948afa20b7(Thread-45 ) WARNING 2018-01-16 14:16:13,877 web:112:web:(log_error) ERROR [192.168.150.101] PUT /images/81569ab3-1b92-4744-8a58-f1948afa20b7: [401] Not authorized (0.00s) GianlucaBTW:I don't know if its is in some way related with the upload problems, but in my engine .log I see these kind of messages every 5 or such seconds:2018-01-16 14:27:38,428+01 INFO [org.ovirt.engine.core.vdsbroker.gluster. GlusterServersListVDSCommand] (DefaultQuartzScheduler3) [61e72c38] START, GlusterServersListVDSCommand( HostName = ovirt02.localdomain.local, VdsIdVDSCommandParametersBase: {hostId='cb9cc605-fceb-4689- ad35-43ba883f4556'}), log id: 65e60794 2018-01-16 14:27:38,858+01 INFO [org.ovirt.engine.core.vdsbroker.gluster. GlusterServersListVDSCommand] (DefaultQuartzScheduler3) [61e72c38] FINISH, GlusterServersListVDSCommand, return: [192.168.150.103/24:CONNECTED, ovirt03.localdomain.local: CONNECTED, ovirt01.localdomain.local: CONNECTED], log id: 65e60794 2018-01-16 14:27:38,867+01 INFO [org.ovirt.engine.core.vdsbroker.gluster. GlusterVolumesListVDSCommand] (DefaultQuartzScheduler3) [61e72c38] START, GlusterVolumesListVDSCommand( HostName = ovirt02.localdomain.local, GlusterVolumesListVDSParameter s:{hostId='cb9cc605-fceb-4689- ad35-43ba883f4556'}), log id: 6e01993d 2018-01-16 14:27:39,221+01 WARN [org.ovirt.engine.core.vdsbroker.gluster. GlusterVolumesListReturn] (DefaultQuartzScheduler3) [61e72c38] Could not associate brick 'ovirt02.localdomain.local:/ gluster/brick1/engine' of volume '6e2bd1d7-9c8e-4c54-9d85- f36e1b871771' with correct network as no gluster network found in cluster '582badbe-0080-0197-013b- 0000000001c6' 2018-01-16 14:27:39,231+01 WARN [org.ovirt.engine.core.vdsbroker.gluster. GlusterVolumesListReturn] (DefaultQuartzScheduler3) [61e72c38] Could not associate brick 'ovirt02.localdomain.local:/ gluster/brick2/data' of volume '2238c6db-48c5-4071-8929- 879cedcf39bf' with correct network as no gluster network found in cluster '582badbe-0080-0197-013b- 0000000001c6' 2018-01-16 14:27:39,253+01 WARN [org.ovirt.engine.core.vdsbroker.gluster. GlusterVolumesListReturn] (DefaultQuartzScheduler3) [61e72c38] Could not associate brick 'ovirt02.localdomain.local:/ gluster/brick4/iso' of volume '28f99f11-3529-43a1-895c- abf1c66884ab' with correct network as no gluster network found in cluster '582badbe-0080-0197-013b- 0000000001c6' 2018-01-16 14:27:39,255+01 INFO [org.ovirt.engine.core.vdsbroker.gluster. GlusterVolumesListVDSCommand] (DefaultQuartzScheduler3) [61e72c38] FINISH, GlusterVolumesListVDSCommand, return: {2238c6db-48c5-4071-8929- 879cedcf39bf=org.ovirt.engine. core.common.businessentities. gluster.GlusterVolumeEntity@ aa6e9a1e, df0ccd1d-5de6-42b8-a163- ec65c3698da3=org.ovirt.engine. core.common.businessentities. gluster.GlusterVolumeEntity@ 31c29088, 6e2bd1d7-9c8e-4c54-9d85- f36e1b871771=org.ovirt.engine. core.common.businessentities. gluster.GlusterVolumeEntity@ ae82860f, 28f99f11-3529-43a1-895c- abf1c66884ab=org.ovirt.engine. core.common.businessentities. gluster.GlusterVolumeEntity@ 1b6a11e5}, log id: 6e01993d Actualy te gluster network seems ok.Eg[root@ovirt01 glusterfs]# gluster volume info dataVolume Name: dataType: ReplicateVolume ID: 2238c6db-48c5-4071-8929-879cedcf39bf Status: StartedSnapshot Count: 0Number of Bricks: 1 x (2 + 1) = 3Transport-type: tcpBricks:Brick1: ovirt01.localdomain.local:/gluster/brick2/data Brick2: ovirt02.localdomain.local:/gluster/brick2/data Brick3: ovirt03.localdomain.local:/gluster/brick2/data (arbiter) Options Reconfigured:performance.strict-o-direct: onnfs.disable: onuser.cifs: offnetwork.ping-timeout: 30cluster.shd-max-threads: 6cluster.shd-wait-qlength: 10000cluster.locking-scheme: granularcluster.data-self-heal-algorithm: full performance.low-prio-threads: 32features.shard-block-size: 512MBfeatures.shard: onstorage.owner-gid: 36storage.owner-uid: 36cluster.server-quorum-type: servercluster.quorum-type: autonetwork.remote-dio: offcluster.eager-lock: enableperformance.stat-prefetch: offperformance.io-cache: offperformance.read-ahead: offperformance.quick-read: offperformance.readdir-ahead: ontransport.address-family: inet[root@ovirt01 glusterfs]#[root@ovirt01 glusterfs]# gluster volume heal data infoBrick ovirt01.localdomain.local:/gluster/brick2/data Status: ConnectedNumber of entries: 0Brick ovirt02.localdomain.local:/gluster/brick2/data Status: ConnectedNumber of entries: 0Brick ovirt03.localdomain.local:/gluster/brick2/data Status: ConnectedNumber of entries: 0[root@ovirt01 glusterfs]#Gianluca