[OPNFLWPLUG-504] flows statistics unstable when 80k flow configured Created: 16/Jun/15 Updated: 27/Sep/21 Resolved: 18/Dec/15 |
|
| Status: | Resolved |
| Project: | OpenFlowPlugin |
| Component/s: | General |
| Affects Version/s: | None |
| Fix Version/s: | None |
| Type: | Bug | ||
| Reporter: | Peter Gubka | Assignee: | Michal Rehak |
| Resolution: | Won't Do | Votes: | 0 |
| Labels: | None | ||
| Remaining Estimate: | Not Specified | ||
| Time Spent: | Not Specified | ||
| Original Estimate: | Not Specified | ||
| Environment: |
Operating System: Linux |
||
| Attachments: |
|
||||||||
| Issue Links: |
|
||||||||
| External issue ID: | 3762 | ||||||||
| Priority: | High | ||||||||
| Description |
|
used odl: distribution-karaf-0.3.0-Lithium-RC1-v201506160017.tar.gz installed feature: odl-openflowplugin-flowservices-ui-li when installed 80k flows, they were not collected, but after mininet restart all flow were collected well then after some time 4 switches were lost from inventory (1434487760.13342, 63, 74567, 74120) this can be just another variant of the same problem, which was reported here |
| Comments |
| Comment by Peter Gubka [ 16/Jun/15 ] |
|
Attachment karaf.log has been added with description: karaf log |
| Comment by Kavitha Ramalingam [ 06/Aug/15 ] |
|
Hi Peter • Is this problem always re-creatable with 63 switches. We hit memory issues on our VMs, so couldn’t try with 63 switches. Thanks and Regards |
| Comment by Kavitha Ramalingam [ 06/Aug/15 ] |
|
Also, i would like to know about how Thanks and regards |
| Comment by Kavitha Ramalingam [ 06/Aug/15 ] |
|
I found that |
| Comment by Kavitha Ramalingam [ 12/Aug/15 ] |
|
I used latest image for my testing. I was able to successfully add 80K flows using the script flow_stats_stability_monitor.py and statistics were in tact. I'll wait for Peter to get back to understand the exact problem recreation steps. Excerpt from the o/p...
(1439379949.190588, 63, 76640, 76000) |
| Comment by Peter Gubka [ 17/Aug/15 ] |
|
The tupple (1434487760.13342, 63, 74567, 74120) contains Since this bug was reported there are jobs like which automate the same workflow as it was done manually before. They use 100k flows and 63 switches. They are robotized. In the job builds #102 and #103 you can see that 100k was not collected with failure(build 103: Also you can see that flow stats collection does not fail all the time. That it why i say "unstable" stats collection. All the builds which have all 11 tests passed were ok. |
| Comment by Kavitha Ramalingam [ 20/Aug/15 ] |
|
I tried running the script more than 10 times. However, i was not able to hit the problem. Looks like its a timing issue. I'll go thro' the code and see if there are any possibility for loophole. Peter, meanwhile can you pl. let me know if i can get access to your test set-up where i could recreate this problem. |
| Comment by Abhijit Kumbhare [ 25/Sep/15 ] |
|
Any thoughts Peter about reproducing this? |
| Comment by Peter Gubka [ 28/Sep/15 ] |
|
I dont know how to reproduce this "on demand". I have been thinking about this since i reported this behavior. But having a look at yellow dots at https://jenkins.opendaylight.org/releng/view/openflowplugin/job/openflowplugin-csit-1node-periodic-scale-stats-collection-daily-lithium-redesign-only-stable-lithium/ we see that it is happening often. |
| Comment by Anil Vishnoi [ 03/Oct/15 ] |
|
Abhijit, this is related to statistics manager in alternate design, so i think michal would be the best person to comment on this. Anil |
| Comment by Abhijit Kumbhare [ 09/Oct/15 ] |
|
Michal, Can you comment? Thanks, |
| Comment by Kamal Rameshan [ 09/Oct/15 ] |
|
Hi Peter , Are you running in a single mininet all the switches? Sometimes i have seen disconnects and flows not getting properly installed in a single mininet vm. Can you retry by distributing the switches amongst more than 1 vm? |
| Comment by Peter Gubka [ 12/Oct/15 ] |
|
(In reply to Kamal Rameshan from comment #11) Hello. |
| Comment by Abhijit Kumbhare [ 30/Oct/15 ] |
|
Made to "normal" as stalled due to no easy way to reproduce the bug. |
| Comment by Peter Gubka [ 18/Dec/15 ] |
|
This bug is covered by the suite The particular test is "Stable State Monitoring". This test has failed in the past only when the previous test failed, because 100k flows were not present in the ds/operational. This bug would be valid only when this particular test fail as the only one in this suite. And this has not happened for several months. This bug was probably fixed as a side effect of stats manager improvement in the past. So closing .... |