<!-- 
RSS generated by JIRA (8.20.10#820010-sha1:ace47f9899e9ee25d7157d59aa17ab06aee30d3d) at Wed Feb 07 19:57:02 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>OpenDaylight JIRA</title>
    <link>https://jira.opendaylight.org</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>8.20.10</version>
        <build-number>820010</build-number>
        <build-date>22-06-2022</build-date>
    </build-info>


<item>
            <title>[CONTROLLER-2032] Phosphorus SR1 - AskTimeoutException while trying to mount multiple Netconf devices at a time</title>
                <link>https://jira.opendaylight.org/browse/CONTROLLER-2032</link>
                <project id="10113" key="CONTROLLER">controller</project>
                    <description>&lt;p&gt;Build Version: &lt;b&gt;Phosphorus SR1&lt;/b&gt;&lt;/p&gt;

&lt;p&gt;3 Node Cluster&lt;/p&gt;

&lt;p&gt;From the Logs - &lt;b&gt;member_1&lt;/b&gt; is chosen as the Master for the Devices&#160;&lt;/p&gt;

&lt;p&gt;Scenario:&lt;/p&gt;

&lt;p&gt;--------------------&lt;/p&gt;

&lt;p&gt;Triggered 200 Mount requests at the same time to ODL.&lt;/p&gt;

&lt;p&gt;In the logs, could see the requests are received and the Mount process was started for few of the devices, but after sometime we start getting the&#160;AskTimeoutException continuously which causes the further device mounts.&lt;/p&gt;

&lt;p&gt;When checked the mount status only around 78 devices among 200 were Connected, and other devices mount got stuck due to above, it dint even span the Netconf actor for the mounting.&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;2022-02-22T20:43:22,747 | ERROR | opendaylight-cluster-data-akka.actor.default-dispatcher-46 | SupervisorStrategy&#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160;| 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Ask timed out on &lt;a href=&quot;#-1603400317&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;Actor[akka://opendaylight-cluster-data/system/typedDdataReplicator#-1603400317]&lt;/a&gt;] after &lt;span class=&quot;error&quot;&gt;&amp;#91;5000 ms&amp;#93;&lt;/span&gt;. Message of type &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Get&amp;#93;&lt;/span&gt;. A typical reason for `AskTimeoutException` is that the recipient actor didn&apos;t send a reply.2022-02-22T20:43:22,747 | ERROR | opendaylight-cluster-data-akka.actor.default-dispatcher-46 | SupervisorStrategy&#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160;| 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Ask timed out on &lt;a href=&quot;#-1603400317&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;Actor[akka://opendaylight-cluster-data/system/typedDdataReplicator#-1603400317]&lt;/a&gt;] after &lt;span class=&quot;error&quot;&gt;&amp;#91;5000 ms&amp;#93;&lt;/span&gt;. Message of type &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Get&amp;#93;&lt;/span&gt;. A typical reason for `AskTimeoutException` is that the recipient actor didn&apos;t send a reply.java.util.concurrent.TimeoutException: Ask timed out on &lt;a href=&quot;#-1603400317&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;Actor[akka://opendaylight-cluster-data/system/typedDdataReplicator#-1603400317]&lt;/a&gt;] after &lt;span class=&quot;error&quot;&gt;&amp;#91;5000 ms&amp;#93;&lt;/span&gt;. Message of type &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Get&amp;#93;&lt;/span&gt;. A typical reason for `AskTimeoutException` is that the recipient actor didn&apos;t send a reply. at akka.actor.typed.scaladsl.AskPattern$.$anonfun$onTimeout$1(AskPattern.scala:131) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.pattern.PromiseActorRef$.$anonfun$apply$1(AskSupport.scala:730) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:479) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at scala.concurrent.ExecutionContext$parasitic$.execute(ExecutionContext.scala:222) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:365) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.executeBucket$1(LightArrayRevolverScheduler.scala:314) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.nextTick(LightArrayRevolverScheduler.scala:318) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.run(LightArrayRevolverScheduler.scala:270) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at java.lang.Thread.run(Thread.java:829) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;?:?&amp;#93;&lt;/span&gt;2022-02-22T20:43:22,788 | INFO&#160; | opendaylight-cluster-data-akka.actor.default-dispatcher-18 | RemoteActorRefProvider$RemoteDeadLetterActorRef | 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Message &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$GetSuccess&amp;#93;&lt;/span&gt; to Actor&lt;span class=&quot;error&quot;&gt;&amp;#91;akka://opendaylight-cluster-data/deadLetters&amp;#93;&lt;/span&gt; was not delivered. &lt;span class=&quot;error&quot;&gt;&amp;#91;5&amp;#93;&lt;/span&gt; dead letters encountered. If this is not an expected behavior then Actor&lt;span class=&quot;error&quot;&gt;&amp;#91;akka://opendaylight-cluster-data/deadLetters&amp;#93;&lt;/span&gt; may have terminated unexpectedly. This logging can be turned off or adjusted with configuration settings &apos;akka.log-dead-letters&apos; and &apos;akka.log-dead-letters-during-shutdown&apos;.2022-02-22T20:43:22,726 | ERROR | opendaylight-cluster-data-akka.actor.default-dispatcher-15 | Behavior$&#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; | 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Supervisor StopSupervisor saw failure: Ask timed out on &lt;a href=&quot;#-1603400317&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;Actor[akka://opendaylight-cluster-data/system/typedDdataReplicator#-1603400317]&lt;/a&gt;] after &lt;span class=&quot;error&quot;&gt;&amp;#91;5000 ms&amp;#93;&lt;/span&gt;. Message of type &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Update&amp;#93;&lt;/span&gt;. A typical reason for `AskTimeoutException` is that the recipient actor didn&apos;t send a reply.java.util.concurrent.TimeoutException: Ask timed out on &lt;a href=&quot;#-1603400317&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;Actor[akka://opendaylight-cluster-data/system/typedDdataReplicator#-1603400317]&lt;/a&gt;] after &lt;span class=&quot;error&quot;&gt;&amp;#91;5000 ms&amp;#93;&lt;/span&gt;. Message of type &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Update&amp;#93;&lt;/span&gt;. A typical reason for `AskTimeoutException` is that the recipient actor didn&apos;t send a reply. at akka.actor.typed.scaladsl.AskPattern$.$anonfun$onTimeout$1(AskPattern.scala:131) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.pattern.PromiseActorRef$.$anonfun$apply$1(AskSupport.scala:730) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:479) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at scala.concurrent.ExecutionContext$parasitic$.execute(ExecutionContext.scala:222) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:365) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.executeBucket$1(LightArrayRevolverScheduler.scala:314) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.nextTick(LightArrayRevolverScheduler.scala:318) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at akka.actor.LightArrayRevolverScheduler$$anon$3.run(LightArrayRevolverScheduler.scala:270) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;bundleFile:?&amp;#93;&lt;/span&gt; at java.lang.Thread.run(Thread.java:829) ~&lt;span class=&quot;error&quot;&gt;&amp;#91;?:?&amp;#93;&lt;/span&gt;2022-02-22T20:43:22,805 | INFO&#160; | opendaylight-cluster-data-akka.actor.default-dispatcher-16 | ClusterSingletonManager&#160; &#160; &#160; &#160; &#160; | 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Singleton actor &lt;span class=&quot;error&quot;&gt;&amp;#91;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor&amp;#93;&lt;/span&gt; was terminated2022-02-22T20:43:22,806 | INFO&#160; | opendaylight-cluster-data-akka.actor.default-dispatcher-16 | LocalActorRef&#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; | 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Message &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Changed&amp;#93;&lt;/span&gt; wrapped in &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.actor.typed.internal.AdaptMessage&amp;#93;&lt;/span&gt; to Actor&lt;a href=&quot;#901657858&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor#901657858&lt;/a&gt; was not delivered. &lt;span class=&quot;error&quot;&gt;&amp;#91;6&amp;#93;&lt;/span&gt; dead letters encountered. If this is not an expected behavior then Actor&lt;a href=&quot;#901657858&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor#901657858&lt;/a&gt; may have terminated unexpectedly. This logging can be turned off or adjusted with configuration settings &apos;akka.log-dead-letters&apos; and &apos;akka.log-dead-letters-during-shutdown&apos;.2022-02-22T20:43:22,807 | INFO&#160; | opendaylight-cluster-data-akka.actor.default-dispatcher-16 | LocalActorRef&#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; | 213 - org.opendaylight.controller.repackaged-akka - 4.0.7 | Message &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.cluster.ddata.typed.javadsl.Replicator$Changed&amp;#93;&lt;/span&gt; wrapped in &lt;span class=&quot;error&quot;&gt;&amp;#91;akka.actor.typed.internal.AdaptMessage&amp;#93;&lt;/span&gt; to Actor&lt;a href=&quot;#901657858&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor#901657858&lt;/a&gt; was not delivered. &lt;span class=&quot;error&quot;&gt;&amp;#91;7&amp;#93;&lt;/span&gt; dead letters encountered. If this is not an expected behavior then Actor&lt;a href=&quot;#901657858&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor#901657858&lt;/a&gt; may have terminated unexpectedly. This logging can be turned off or adjusted with configuration settings &apos;akka.log-dead-letters&apos; and &apos;akka.log-dead-letters-during-shutdown&apos;.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</description>
                <environment></environment>
        <key id="35252">CONTROLLER-2032</key>
            <summary>Phosphorus SR1 - AskTimeoutException while trying to mount multiple Netconf devices at a time</summary>
                <type id="10104" iconUrl="https://jira.opendaylight.org/secure/viewavatar?size=xsmall&amp;avatarId=10303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.opendaylight.org/images/icons/priorities/blocker.svg">Highest</priority>
                        <status id="3" iconUrl="https://jira.opendaylight.org/images/icons/statuses/inprogress.png" description="This issue is being actively worked on at the moment by the assignee.">In Progress</status>
                    <statusCategory id="4" key="indeterminate" colorName="yellow"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="shibu.vijayakumar">Shibu Vijayakumar</assignee>
                                    <reporter username="shibu.vijayakumar">Shibu Vijayakumar</reporter>
                        <labels>
                    </labels>
                <created>Tue, 22 Feb 2022 21:18:18 +0000</created>
                <updated>Wed, 6 Apr 2022 17:14:15 +0000</updated>
                                            <version>4.0.10</version>
                                                    <component>clustering</component>
                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                                                                <comments>
                            <comment id="70556" author="JIRAUSER14601" created="Tue, 22 Feb 2022 21:23:05 +0000"  >&lt;p&gt;&lt;b&gt;Major Concern on this Issue is:&lt;/b&gt; Once we hit this timeout issue, no further mount request is processed by ODL.&lt;/p&gt;

&lt;p&gt;Tried giving a new POST mount request for a new device other than the above 200 - device vmx-201, api was passed, but the actual mount process was not started by the ODL.&#160;&lt;/p&gt;

&lt;p&gt;Even after trying Deleting the 200 devices and then trying a new mount dint go through. But delete process was successful.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Device 201 Single mount after the 200 parallel mount was performed at&#160;&lt;br/&gt;
&lt;b&gt;2022-02-22T20:48:14,434&lt;/b&gt; | INFO | opendaylight-cluster-data-notification-dispatcher-59 | DefaultSchemaResourceManager | 287 - org.opendaylight.netconf.sal-netconf-connector - 2.0.11 | Netconf connector for device RemoteDevice{vmx-201} will use schema cache directory junos instead of schema&lt;/p&gt;

&lt;p&gt;From the logs it could see, DOMEntity of the&#160;EntityTypeListenerActor as&#160;&lt;/p&gt;

&lt;p&gt;_nodeId=Uri{_value=vmx-201}}]]}}]], state=REMOTE_OWNERSHIP_LOST_NO_OWNER &lt;span class=&quot;error&quot;&gt;&amp;#91;wasOwner=false, isOwner=false, hasOwner=false&amp;#93;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="70557" author="JIRAUSER14601" created="Wed, 23 Feb 2022 14:31:18 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.opendaylight.org/secure/ViewProfile.jspa?name=rovarga&quot; class=&quot;user-hover&quot; rel=&quot;rovarga&quot;&gt;rovarga&lt;/a&gt;&#160;Another observation, once we get this error in the ODL instance which has the Master role was the devices mount, if we down that instance - could see another election happens and one of the other ODL gets the Leader role and after that it resumes all the above mounts that were stuck and all devices moves to Connected.&lt;/p&gt;</comment>
                            <comment id="70621" author="rovarga" created="Thu, 3 Mar 2022 16:02:40 +0000"  >&lt;p&gt;So the ATE is troublesome, as it seems to indicate something in Akka replication. Two things:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;try to monitor the VMs/JVMs for signs of issues, most notably GC pauses and similar&lt;/li&gt;
	&lt;li&gt;try the test case with odl-netconf-topology, i.e. without cluster, to bring down the amount of moving pieces&lt;/li&gt;
&lt;/ol&gt;
</comment>
                            <comment id="70622" author="JIRAUSER14501" created="Thu, 3 Mar 2022 19:55:02 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.opendaylight.org/secure/ViewProfile.jspa?name=rovarga&quot; class=&quot;user-hover&quot; rel=&quot;rovarga&quot;&gt;rovarga&lt;/a&gt;&#160;Sure, will check on the point 1.&lt;/p&gt;

&lt;p&gt;W.r.t point 2 , have executed the test case and it was SUCCESS with no ATE in single instance ODL, enabling&#160;odl-netconf-topology, also in the Clustered ODL setup of 3 node, with the feature:&#160;odl-netconf-topology installed instead of &quot;odl-netconf-clustered-topology&quot;.&lt;/p&gt;

&lt;p&gt;But the concern with the 3 node clustered ODL in the above case is the Netconf device Mount Point is created only in one of the instance where the POST mount api hits, and in the other 2 its not, and due to that the below GET and PUT netconf apis were failing in those instances, saying &quot;mount point doesn&apos;t exists&quot;.&#160;&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;em&gt;Any workaround or suggestions on this scenario to make it work as a Cluster with the&lt;/em&gt;&lt;/b&gt;&#160;&lt;br/&gt;
 &lt;b&gt;&lt;em&gt;odl-netconf-topology feature installed?&lt;/em&gt;&lt;/b&gt;&lt;/p&gt;

&lt;p&gt;Sample apis:&lt;/p&gt;

&lt;p&gt;&lt;b&gt;GET&lt;/b&gt;:&#160;curl -v GET -u admin:admin http://&amp;lt;ip&amp;gt;:8181/restconf/operational/network-topology:network-topology/topology/topology-netconf/node/&amp;lt;device_name&amp;gt;/yang-ext:mount&lt;/p&gt;

&lt;p&gt;&lt;b&gt;PUT&lt;/b&gt;:&#160;curl -H &apos;Content-Type: application/json&apos; -H &apos;Accept: application/json&apos; -v -X PUT -T policy-options.json -u admin:admin &lt;a href=&quot;http://10.100.100.105:30181/restconf/config/network-topology:network-topology/topology/topology-netconf/node/18579-sim-device/yang-ext:mount/junos-conf-root:configuration/junos-conf-policy-options:policy-options&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://&amp;lt;ip&amp;gt;:8181/restconf/config/network-topology:network-topology/topology/topology-netconf/node/&amp;lt;device_name&amp;gt;/yang-ext:mount/junos-conf-root:configuration/junos-conf-policy-options:policy-options&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="70623" author="rovarga" created="Thu, 3 Mar 2022 20:57:14 +0000"  >&lt;p&gt;Right, and I would expect single-node deployments to not be suspectible to timing (quoted 5000ms). Distributed systems are pesky &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.opendaylight.org/images/icons/emoticons/sad.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;

&lt;p&gt;In a cluster, odl-netconf-topology has a snowflake&apos;s chance in hell to work &#8211; the three instances will end up tripping over each other.&lt;/p&gt;

&lt;p&gt;But, if that works okay, it is a matter of clustering integration and we need to understand what is going wrong. Unfortunately I cannot look at this closely enough because of other (long overdue) stuff...&lt;/p&gt;</comment>
                            <comment id="70625" author="JIRAUSER14601" created="Mon, 7 Mar 2022 09:44:36 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.opendaylight.org/secure/ViewProfile.jspa?name=rovarga&quot; class=&quot;user-hover&quot; rel=&quot;rovarga&quot;&gt;rovarga&lt;/a&gt;&#160;In the above point, u r suggesting we could check and enhance &quot;odl-netconf-topology&quot; to work as a cluster to create &quot;Mount points&quot; in all of the ODL instance?&lt;/p&gt;</comment>
                            <comment id="70630" author="JIRAUSER14601" created="Wed, 9 Mar 2022 22:02:13 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.opendaylight.org/secure/ViewProfile.jspa?name=rovarga&quot; class=&quot;user-hover&quot; rel=&quot;rovarga&quot;&gt;rovarga&lt;/a&gt;&#160;I hope I kind of have figured out the root cause for the above mount stuck issue,&#160;&lt;/p&gt;

&lt;p&gt;&lt;b&gt;Reason&lt;/b&gt;:&#160;Singleton actor &lt;span class=&quot;error&quot;&gt;&amp;#91;akka://opendaylight-cluster-data/system/singletonManagerOwnerSupervisor/OwnerSupervisor&amp;#93;&lt;/span&gt;&#160;was terminated&lt;/p&gt;

&lt;p&gt;And this actor doesn&apos;t resumes. Ideally the ClusterSingletonManager should make sure that this Single actor always runs right? Even if it terminates it should be restarted I think.&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;I could understand controller/&lt;b&gt;EOSMain.java&lt;/b&gt; is the one creating this ClusterSingleton actor,&lt;/p&gt;

&lt;p&gt;// start the initial sync behavior that switches to the regular one after syncing&lt;br/&gt;
 ownerSupervisor = clusterSingleton.init(&lt;br/&gt;
 SingletonActor.of(IdleSupervisor.create(iidCodec), &quot;OwnerSupervisor&quot;));&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;And the class of this actor is controller/&lt;b&gt;OwnerSupervisor.java&lt;/b&gt;,&#160;&lt;/p&gt;

&lt;p&gt;as this actor is terminated, the DOMEntities are stuck in a REMOTE_OWNERSHIP_LOST_NO_OWNER stage and no odl instance takes the master ownership (LOCAL_OWNERSHIP_GRANTED) and proceeds with the mount.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;If we do restart the ODL instance, then this Singleton&#160;OwnerSupervisor is getting started in any of the other ODL instance and that resumes the on-hold mounts.&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Could see&#160;akka.cluster.singleton.hand-over-retry-interval&#160;= 1s is configured in factory/akka.conf, but still not sure why the OwnerSupervisor actor doesn&apos;t restarts in the other instance when it gets terminated.&lt;/p&gt;</comment>
                            <comment id="70707" author="JIRAUSER14601" created="Mon, 4 Apr 2022 07:02:04 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.opendaylight.org/secure/ViewProfile.jspa?name=rovarga&quot; class=&quot;user-hover&quot; rel=&quot;rovarga&quot;&gt;rovarga&lt;/a&gt;&#160;Found a solution for the above issue, when increased the below &quot;distributed-data&quot; intervals from ms to 5s in akka.conf, we were able to mount 250 netconf devices at the same time to the ODL.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;akka.cluster.distributed-data.gossip-interval = 5s&lt;/p&gt;

&lt;p&gt;akka.cluster.distributed-data.notify-subscribers-interval = 5s&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;akka {&lt;/p&gt;

&lt;p&gt;&#160;&#160;&#160;&#160; cluster {&lt;/p&gt;

&lt;p&gt;&#160;&#160;&#160;&#160;&#160;&#160;&#160; distributed-data &lt;/p&gt;
{

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # How often the Replicator should send out gossip information.

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # This value controls how quickly Entity Ownership Service data is replicated

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # across cluster nodes.

&#160;&#160;&#160;&#160;&#160;&#160;&#160; gossip-interval = 5s

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # How often the subscribers will be notified of changes, if any.

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # This value controls how quickly Entity Ownership Service decisions are

&#160;&#160;&#160;&#160;&#160;&#160;&#160; # propagated within a node.

&#160;&#160;&#160;&#160;&#160;&#160;&#160; notify-subscribers-interval = 5s

&#160;&#160;&#160;&#160;&#160; }

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="70708" author="JIRAUSER14601" created="Mon, 4 Apr 2022 07:06:11 +0000"  >&lt;p&gt;Added fix for restarting the EOS Singleton Actor -&#160;&lt;a href=&quot;https://jira.opendaylight.org/browse/CONTROLLER-2035&quot; title=&quot;ODL Cluster - Akka Cluster Singleton Manager Actor - OwnerSupervisor Getting Terminated and never restarts&quot; class=&quot;issue-link&quot; data-issue-key=&quot;CONTROLLER-2035&quot;&gt;&lt;del&gt;CONTROLLER-2035&lt;/del&gt;&lt;/a&gt; ODL Cluster - Akka Cluster Singleton Manager Actor - OwnerSupervisor Getting Terminated and never restarts - OpenDaylight JIRA&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="17361" name="member1_karaf.log" size="4097421" author="shibu.vijayakumar" created="Tue, 22 Feb 2022 21:17:54 +0000"/>
                            <attachment id="17360" name="member2_karaf.log" size="3666671" author="shibu.vijayakumar" created="Tue, 22 Feb 2022 21:17:54 +0000"/>
                            <attachment id="17359" name="member3_karaf.log" size="3643388" author="shibu.vijayakumar" created="Tue, 22 Feb 2022 21:17:54 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                            <customfield id="customfield_11400" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                <customfield id="customfield_10000" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>0|hzzzxw:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                </customfields>
    </item>
</channel>
</rss>