15:03:26 #startmeeting neutron_northbound 15:03:26 Meeting started Mon Mar 27 15:03:26 2017 UTC. The chair is yamahata. Information about MeetBot at http://ci.openstack.org/meetbot.html. 15:03:26 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:03:26 The meeting name has been set to 'neutron_northbound' 15:03:29 sridharg : ping 15:03:36 #topic agenda bashing and roll call 15:03:41 #info yamahata 15:03:49 #info reedip 15:03:52 #link https://wiki.opendaylight.org/view/NeutronNorthbound:Meetings meeting agend 15:04:16 in addition to usual topics, any other topic to discuss today? 15:04:27 #info rajivk 15:04:44 yamahata : IPv6 ICMP :) 15:04:45 From last week, patches to be discussed and we have several new patches/bug reports. 15:04:57 icmpv6, yeah. 15:06:00 #topic Announcements 15:06:05 any announcement? 15:06:22 ODL Carbon is near to release. 15:06:52 ODL neutron northbound is code freeze. Karaf4 needs one more patch so that integration test is run with karaf4 15:07:10 ODL DDF is announced officially 15:07:44 #link http://events.linuxfoundation.org/events/opendaylight-developer-design-forum ODL ddf 15:08:12 For now, only place and date is announce. the details isn't announced and the registration is not open yet. 15:08:20 Santa Clara, CA 15:08:27 May 31- June 2, 2017 15:08:42 From me, that's it. any other announcement? 15:09:32 seems nothing. move on. 15:09:34 #topic action items from last meeting 15:09:54 only patch review and bug scrub. 15:10:08 #topic Pike planning 15:10:16 We have two spec 15:10:24 in review process 15:10:47 #link https://review.openstack.org/#/c/443541/ 15:10:47 Propose spec for dependency validations move 15:10:57 This one is discussed last week 15:11:11 #link https://review.openstack.org/#/c/449402/ spec: rpc from odl to networking-odl 15:11:16 This one is new one I've uploaded 15:12:01 regarding to 443541. Only issue is performance/benchmark. Otherwise there is consensus. 15:12:13 Probably we can merge it after minor review resping. 15:12:22 Regarding to 449402, this is new one I've uploaded. 15:13:15 Probably we need to discuss with actual user. i.e. netvirt dhcp people. Unfortunately we don't have them today. 15:13:33 hi sorry im late 15:13:39 mkolesni: hello 15:13:47 Now we're discussing pike planning. 15:14:00 we have two spec uploaded under review. 15:14:23 #link https://blueprints.launchpad.net/networking-odl blueprint of networking-odl 15:14:58 brb 15:14:58 #topic carbon/nitrogen planning 15:15:31 for nitrogen, we need to create release plan soon. 15:16:00 Nothing urgent 15:16:13 #topic bugs/patches 15:16:20 Okay. Now discussion time. 15:16:31 There are several bug reports and tempest failures. 15:16:52 Especially tempest failures are blocking patch backports. 15:17:14 Is anyone looking into it? 15:17:36 yamahata, do you mean Alon patches 15:17:44 yamahata: is there a grafanna dashboard ? 15:17:51 rajivk: yes for backporting. 15:18:07 i havent had a chance to look yet 15:18:13 #link http://grafana.openstack.org/dashboard/db/networking-odl-failure-rate grafana for netowkring-odl 15:18:25 is everything on stable broken? 15:18:26 yeah, i checked them. 15:18:28 master seems fine 15:18:38 Yes, even it's broken for master branch. 15:18:41 yeah got it yamahata 15:18:48 I will investigate it. 15:18:49 wow, py27 is broken 15:18:52 As temporal work around, several test cases are disabled 15:18:57 rajivk +1 15:19:21 yamahata, i want clarity on it. 15:19:47 rajivk: sure, what needs clarity? 15:19:54 You said master is broken, i tested today. 15:19:56 yamahata: only v1 driver tempest is failing 15:19:59 It installed successfully. 15:20:18 ohh, you mean tempest tests are failing. 15:21:02 Regarding to v1 driver, probably we can ignore/disable it as v1driver is being depricated. 15:21:03 rajivk : py27 and tempest both are failing Currently 15:21:09 thats weird since the change is for v2 15:21:18 yamahata, im not sure we should 15:21:28 v2 is default only from ocata 15:21:37 yamahata as per grafanna, v1 driver is failing, so we can disable it and then check the tempest status .. what do you think? 15:21:57 mkolesni: makes sense. 15:21:59 mkolesni, +1 15:22:22 seems like some infra issue there 15:22:25 mkolesni : v1 can disabled for pike, isnt it? 15:22:32 tests didnt even run 15:22:49 Did we file a bug report for it? 15:22:51 reedip, sure i guess 15:23:00 mkolesni, Yamahata 15:23:06 How many tests are failing? 15:23:16 rajivk, they didnt even run 15:23:21 the infra failed 15:23:30 rajivk , you can also check http://grafana.openstack.org/dashboard/db/networking-odl-failure-rate 15:23:58 2017-03-24 05:12:54.044550 | + /opt/stack/new/devstack-gate/devstack-vm-gate.sh:main:L719: /tmp/ansible/bin/ansible primary -f 5 -i /home/jenkins/workspace/gate-tempest-dsvm-networking-odl-carbon-snapshot-v1driver/inventory -m shell -a 'cd '\''/opt/stack/new/devstack'\'' && sudo -H -u stack DSTOOLS_VERSION=0.3.0 stdbuf -oL -eL ./stack.sh executable=/bin/bash' 15:23:58 2017-03-24 05:12:58.046981 | ERROR: the main setup script run by this job failed - exit code: 2 15:24:22 mkolensi: this one ?? http://logs.openstack.org/27/449727/1/check/gate-grenade-dsvm-networking-odl-nv/c10a925/console.html#_2017-03-24_17_24_53_896745 15:24:52 i think, i checked it. 15:25:03 let me verify 15:25:16 im looking here http://logs.openstack.org/75/448475/5/check/gate-tempest-dsvm-networking-odl-carbon-snapshot-v1driver/59bc7fd/logs/ 15:25:38 which file mkolesni ? 15:25:55 look in the console log 15:26:06 also test_results.html is empty 15:26:11 its for this change https://review.openstack.org/#/c/448475/ 15:26:27 http://logs.openstack.org/75/448475/5/check/gate-tempest-dsvm-networking-odl-carbon-snapshot-v1driver/59bc7fd/logs/devstack-early.txt.gz 15:26:57 mkolesni, it is for newton branch. 15:27:02 the time is only several minutes. So it failed to setup running environment. 15:27:20 Is master also failing with the same error? 15:27:21 hi all, sorry for being late. 15:27:33 hi vthapar 15:27:42 vthapar: hello 15:27:50 probably some change in the dependencies of the stable branch caused this 15:29:21 rajivk : how did you determine its newton ?? 15:29:39 any suggestions? 15:30:18 the link mkolesni shared, it is in stable/newton 15:31:08 oh , but I looked at the recent logs as well buddy 15:31:27 please share the link 15:32:08 the devstack-early log show that devstack is wating for password for rabbitmq. 15:32:19 the config seems changed. 15:32:32 yamahata, link please 15:32:50 yamahata : yep 15:32:50 Oh no. I was wrong? 15:33:01 http://logs.openstack.org/75/448475/5/check/gate-tempest-dsvm-networking-odl-carbon-snapshot-v1driver/59bc7fd/logs/devstack-early.txt.gz 15:33:04 your link. 15:33:08 No I found that as well in mkolensi's log 15:34:10 Do we want to continue debug session? or file a bug and move on to other bugs/patches? 15:34:23 anyway I'll volunteer to file a bug 15:34:34 #action yamahata file a bug for stable branch breakage. 15:34:52 yes , kets do it after the meeting 15:35:06 yamahata, the discussion we were having, was it for newton or master? 15:36:08 I've thought for both. 15:36:14 master seems to be fine, i submitted a doc patch 15:36:19 It was fine. 15:36:32 does doc patch does not have functional and fulltest? 15:36:34 https://review.openstack.org/#/c/446390/ 15:36:41 Ah devstack is okay. there is ssh failure for master 15:36:56 some test cases are disabled for master. 15:37:09 rajivk functional is working 15:37:13 devstack fails for stable(ocata, newton) branch 15:37:17 http://logs.openstack.org/90/446390/5/check/gate-dsvm-networking-odl-functional-carbon-snapshot/a8618d2/ 15:37:44 but I think fullstack is not there 15:37:48 So it means, master as of now correct, because some of the test cases are disabled. 15:38:13 yamahata, If we enabled them then master will also fail, is my understanding correct? 15:38:21 rajivk: right. 15:38:43 yamahata, you are talking about https://review.openstack.org/#/c/445251/ 15:38:48 https://review.openstack.org/#/c/445251/ doesn't pass 15:38:54 rajivk: yes. 15:39:09 I checked it and put a comment. 15:39:26 May be you can provide your opinion on it. 15:39:49 Did you check my comment. Let's leave this discussion for now. 15:40:20 I haven't checked it deeply. I'll check it. 15:40:41 #action yamahata check rajivk comment on https://review.openstack.org/#/c/445251/ 15:41:25 Now can we move on to other bugs/patches? 15:41:37 yes, we can :) 15:41:49 can i get reviews on remaining patches on https://review.openstack.org/#/q/status:open+project:openstack/networking-odl+branch:master+topic:journal-singleton please? 15:41:55 #link https://review.openstack.org/#/q/status:open+project:openstack/networking-odl+branch:master+topic:journal-singleton 15:42:19 i fixed it according to the comments 15:42:28 Cool. 15:42:50 anything else? icmpv6? 15:43:05 wait 15:43:17 reedip? 15:43:22 hi 15:44:03 #link https://review.openstack.org/#/c/449825/ 15:44:03 secgroup: convert icmpv6 variant name into icmpv6 15:44:20 This patch is to response to neutron discussion at mailing list. 15:44:40 icmp and ipv6 can be divided into 3 combinations 15:44:44 icmpv6 15:44:51 ipv6-icmp 15:44:54 and icmp ONLY 15:45:04 It would take a while to fix neutron upstream. But I don't want to leave networking-odl broken until neutron fix. 15:45:10 currently neutron supports all 3 options 15:45:48 When was ipv6-icmp introduced? 15:45:56 From Pike or older? 15:46:00 but it results into duplicate results 15:46:04 Mitaka I guess 15:46:25 Oh, then we may need fix for newton and ocata. 15:46:43 yamahata, wait, let me make it sure 15:48:06 meanwhile I just want tell you guys 15:48:19 that due to multiple values of Ipv6 in SG 15:48:25 duplicate rules are present 15:48:34 and that is an inconvineance 15:48:44 to fix that , I proposed a patch in Neutron 15:48:53 OVS support ipv6-icmp 15:48:59 but ODL supports ICMPv6 15:49:12 so we need to make sure the behavior remains the same 15:49:32 https://review.openstack.org/#/c/259037/ Add popular IP protocols for security group 15:49:50 the patch was merged Feb 5, 2016. 15:50:07 newton and ocata branch needs fix. 15:50:30 for Brian from Neutron ( haleyb) mentioned that Neutron would by default add ipv6-icmp 15:50:42 and as per sridharg, we need to add icmpv6 for ODL 15:51:22 ODL neutron northbound knows icmpv6, but doesn't ipv6-icmpv6. 15:54:00 yamahata : ok , thats what sridarg mentioned 15:54:14 yamahata, mkolesni, I would like to discussion singleton patch from mkolesni. 15:54:39 rajivk: sure. please go ahead. 15:54:55 mkolesni, You mean due to GIL only one thread should be run? 15:54:56 I left a message on https://review.openstack.org/#/c/449825/ 15:55:22 only one thread is run at any given moment 15:55:33 so no need to have more than one doing journaling 15:55:44 hmm, i agree. 15:56:07 yamahata, now i would like to discuss about threadpool 15:56:12 lets say if odl is stuck responding then it doesnt help that we have 10 threads 15:56:25 since they will all be stuck 15:56:37 so it doesnt really save us anything 15:56:47 as mentioned by yamahata, do we still need threadpool for thread management? 15:56:49 In some cases, ODL MD-SAL takes a long time. 15:56:57 In that case, multiple thread helps. 15:57:05 and i/o is the only time afaik the gil is released other than plain context switch 15:57:22 That's right. 15:57:30 if we hit it we can always add thread pool 15:57:33 and http reuqest is IO. db query is io. 15:57:43 but anyway in real environment youll have multiple api workers 15:57:55 so each process should have 1 thread no more 15:58:08 since you can easily get 10's of simultaneous processes 15:58:11 we can have multiple thread. 15:58:25 Why do we have to stick to 1 thread per 1 process? 15:58:43 10thread per 1 process would be much cheaper. 15:58:44 again because of gil there is no real multi threading 15:59:07 gil isn't related. 15:59:16 of course it is 15:59:19 eventlet is greentread and pymysql is full python implementation. 15:59:30 i say we need currently the sim-plest solution 15:59:34 the eventlet thread is switched at io. 15:59:52 and later well see if it has bottlenecks we should solve them 15:59:58 solution for what issue? 16:00:07 for code design 16:00:23 as code is more complicated it is harder to maintain 16:01:19 right now if i look at any log i see 10s of messages 'nothing in journal' popping up all the time 16:01:29 and multiple of them coming from the same process 16:01:40 very hard to debug such logs 16:01:49 and its just noise not really helping 16:02:05 so if we have singleton everything is much simpler 16:02:11 I agree that log is mess. 16:02:20 and later on we can add thread pool if we see theres a bottleneck there 16:02:32 That's because right now journal threads aren't maintained properly. 16:02:36 but right now the number of threads is the number of v2 drivers loaded 16:02:53 right so were basically agreeing here 16:02:53 In that case, thread is much easier to maintain than multiple processes. 16:03:03 yamahata : I saw your comment in sridarg's patch on https://review.openstack.org/#/c/449825/ 16:03:08 but multiple processes happens anyway 16:03:14 So that we can implement smart wakeup mechanism. 16:03:22 instead of waking up all. 16:03:27 you cant avoid it 16:04:01 For example, we can introduce a config for the number of threads. 16:04:12 ok this i think should be honored by a spec that we can review the design and comment on it 16:04:57 Hmm, it's 4 min over. 16:05:13 we can continue discussion on the patch or spec. 16:05:20 do we have any other urgent bugs/patches? 16:05:51 not urgent but long 16:05:57 rajivk: which one? 16:05:59 i gtg 16:06:04 standing patch, eliminate network topology 16:06:08 if nothing else needed from me 16:06:19 bye guys 16:06:25 mkolesni: thanks 16:06:35 mkolesni, thanks have a nice day. 16:06:45 thanks you too :) 16:06:54 rajivk: the patch itself is getting in good shape. 16:07:00 yamahata : I agree with your comment on https://review.openstack.org/#/c/449825/ 16:07:29 so the question is when to merge. Pike or Queens 16:08:04 do you mean network-topogy one? 16:08:17 rajivk: yes. 16:08:38 https://review.openstack.org/#/c/334250/ 16:09:16 yamahata, i want to know more about thread leader election for houskeeping tasks. 16:09:43 what are the tasks that we would taking into consideration. 16:10:11 hostconfig update. 16:10:16 and port status update 16:10:36 currently, i am thinking of using leader election by assigning numbers to each process at startup and later on using rpc to broadcast its heartbeat 16:10:43 tasks that receive info from ODL and update neutron db 16:11:02 we don't want to use rpc between neutron servers. 16:11:09 and then the process having least number will be elected as leader, it is sort of bully algorithm. 16:11:15 We'd like to use db for polling. 16:11:31 something like maintenance thread. 16:11:50 but what do we poll in database? 16:11:58 Yes. polling 16:12:05 we are already moving to make networking-odl as agentless. 16:12:21 the leader election can be slow. we don't need fast take over. 16:12:26 and just poll db. 16:12:37 ok, what do we poll from db? 16:12:52 on what information, we elect leader? 16:13:13 OpendaylightMaintenance 16:13:21 or new db table. 16:13:44 You mean, we create a new db table for maintaince thread leader election process? 16:14:03 yes. Probably we can OpendaylightMaintenance 16:14:12 we can use OpendaylightMaintenance 16:15:18 you mean https://github.com/openstack/networking-odl/blob/master/networking_odl/db/models.py#L45 this one? 16:15:32 yes 16:15:39 Now it's dedicated for maintenance thread. 16:15:56 Probably we can add one column like task or something 16:16:19 okay, i will check code from this perspective. 16:16:49 BTW i would like to know, why don't you want to use rpc? 16:17:05 So far we haven't used rpc. 16:17:21 we don't want to rely on extra mechanism. 16:17:35 and we have already OpendaylightMaintenance 16:17:46 But all the services rely on messsaging service 16:18:14 they communicate through messaging then why not? 16:18:35 Technically we can. 16:19:00 But there must be some factors because of which we prefer database but not messaging? 16:19:06 what are those factors? 16:20:43 Anyway after rpc, we need to keep track of their liveness. 16:20:50 something similar of agentdb 16:21:13 It's OpendaylightMaintenance 16:21:18 actually we need to broadcast messages only for aliveness. 16:21:25 without rpc, each neutron server can directly update the table. 16:21:49 Yes. 16:22:38 we don't need strict leader election. only easy implemetation is wanted. 16:22:53 okay 16:23:31 so I've thought we can just slightly enhance the existing maintenance thread one. 16:23:34 I will consider the table you mentioned. 16:24:31 This meeting is not ended yet :) 16:24:39 oh let's end it 16:24:45 #topic open mike 16:24:46 ok 16:24:51 #topic cookies 16:24:55 #endmeeting