[ovirt-users] CentOS 7 + oVirt 3.5 + OpenVPN
Darrell Budic
budic at onholyground.com
Sat Oct 18 23:04:31 EDT 2014
Looks like an openvpn config issue and not a ovirt issue from this. 192.168.124.1 is not in the same network as 192.168.124.200/25 mostly, try 192.168.124.129.
> On Oct 18, 2014, at 7:45 AM, Phil Daws <uxbod at splatnix.net> wrote:
>
> Hello:
>
> have installed oVirt 3.5 VDSM on a CentOS 7 node and then OpenVPN. The problem I have is that when I start OpenVPN I receive the message:
>
> Oct 18 13:29:50 kvm01 openvpn[4159]: /usr/sbin/ip link set dev tun0 up mtu 1500
> Oct 18 13:29:50 kvm01 openvpn[4159]: /usr/sbin/ip addr add dev tun0 192.168.124.200/25 broadcast 192.168.124.255
> Oct 18 13:29:50 kvm01 openvpn[4159]: /usr/sbin/ip route add 192.168.0.0/16 via 192.168.124.1
> Oct 18 13:29:50 kvm01 openvpn[4159]: ERROR: Linux route add command failed: external program exited with error status: 2
>
> and if I run the route command manually:
>
> [root at kvm01 sysconfig]# /usr/sbin/ip route add 192.168.0.0/16 via 192.168.124.1
> RTNETLINK answers: No such process
>
> It would appear the tunnel is up:
>
> [root at kvm01 sysconfig]# ip add ls
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> inet 127.0.0.1/8 scope host lo
> valid_lft forever preferred_lft forever
> inet6 ::1/128 scope host
> valid_lft forever preferred_lft forever
> 2: bond0: <BROADCAST,MULTICAST,MASTER> mtu 1500 qdisc noop state DOWN
> link/ether f2:c9:ce:e5:ac:32 brd ff:ff:ff:ff:ff:ff
> 3: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master ovirtmgmt state UP qlen 1000
> link/ether c8:1f:66:c4:2c:76 brd ff:ff:ff:ff:ff:ff
> inet6 fe80::ca1f:66ff:fec4:2c76/64 scope link
> valid_lft forever preferred_lft forever
> 4: em2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
> link/ether c8:1f:66:c4:2c:77 brd ff:ff:ff:ff:ff:ff
> 6: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> link/ether 46:af:6e:9a:1e:4b brd ff:ff:ff:ff:ff:ff
> 8: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
> link/ether c8:1f:66:c4:2c:76 brd ff:ff:ff:ff:ff:ff
> inet XXX.XXX.XXX.XXX/23 brd 88.150.253.255 scope global ovirtmgmt
> valid_lft forever preferred_lft forever
> inet6 fe80::ca1f:66ff:fec4:2c76/64 scope link
> valid_lft forever preferred_lft forever
> 10: tun0: <POINTOPOINT,MULTICAST,NOARP,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UNKNOWN qlen 100
> link/none
> inet 192.168.124.200/25 brd 192.168.124.255 scope global tun0
> valid_lft forever preferred_lft forever
>
> Any thoughts as to why the route will not work ? Rationale for this approach is its a cloud server and wish to use a private network to reach the install VMs on that node.
>
> Thanks, Phil
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
More information about the Users
mailing list