org.apache.solr.cloud.ZkController.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.solr.cloud.ZkController.java

Source

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.solr.cloud;

import java.io.IOException;
import java.io.UnsupportedEncodingException;
import java.lang.invoke.MethodHandles;
import java.net.InetAddress;
import java.net.NetworkInterface;
import java.net.URLEncoder;
import java.net.UnknownHostException;
import java.nio.charset.StandardCharsets;
import java.nio.file.Path;
import java.util.ArrayList;
import java.util.Collection;
import java.util.Collections;
import java.util.Enumeration;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Locale;
import java.util.Map;
import java.util.Set;
import java.util.concurrent.Callable;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Future;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.TimeoutException;
import java.util.concurrent.atomic.AtomicReference;

import com.google.common.base.Strings;
import org.apache.commons.lang.StringUtils;
import org.apache.solr.client.solrj.impl.HttpSolrClient;
import org.apache.solr.client.solrj.impl.HttpSolrClient.Builder;
import org.apache.solr.client.solrj.request.CoreAdminRequest.WaitForState;
import org.apache.solr.cloud.overseer.OverseerAction;
import org.apache.solr.cloud.overseer.SliceMutator;
import org.apache.solr.common.SolrException;
import org.apache.solr.common.SolrException.ErrorCode;
import org.apache.solr.common.cloud.BeforeReconnect;
import org.apache.solr.common.cloud.ClusterState;
import org.apache.solr.common.cloud.ClusterStateUtil;
import org.apache.solr.common.cloud.DefaultConnectionStrategy;
import org.apache.solr.common.cloud.DefaultZkACLProvider;
import org.apache.solr.common.cloud.DefaultZkCredentialsProvider;
import org.apache.solr.common.cloud.DocCollection;
import org.apache.solr.common.cloud.OnReconnect;
import org.apache.solr.common.cloud.Replica;
import org.apache.solr.common.cloud.Slice;
import org.apache.solr.common.cloud.SolrZkClient;
import org.apache.solr.common.cloud.ZkACLProvider;
import org.apache.solr.common.cloud.ZkCmdExecutor;
import org.apache.solr.common.cloud.ZkConfigManager;
import org.apache.solr.common.cloud.ZkCoreNodeProps;
import org.apache.solr.common.cloud.ZkCredentialsProvider;
import org.apache.solr.common.cloud.ZkNodeProps;
import org.apache.solr.common.cloud.ZkStateReader;
import org.apache.solr.common.cloud.ZooKeeperException;
import org.apache.solr.common.params.CollectionParams;
import org.apache.solr.common.params.CommonParams;
import org.apache.solr.common.params.SolrParams;
import org.apache.solr.common.util.StrUtils;
import org.apache.solr.common.util.URLUtil;
import org.apache.solr.common.util.Utils;
import org.apache.solr.core.CloseHook;
import org.apache.solr.core.CloudConfig;
import org.apache.solr.core.CoreContainer;
import org.apache.solr.core.CoreDescriptor;
import org.apache.solr.core.SolrCore;
import org.apache.solr.logging.MDCLoggingContext;
import org.apache.solr.update.UpdateLog;
import org.apache.zookeeper.CreateMode;
import org.apache.zookeeper.KeeperException;
import org.apache.zookeeper.KeeperException.ConnectionLossException;
import org.apache.zookeeper.KeeperException.NoNodeException;
import org.apache.zookeeper.KeeperException.SessionExpiredException;
import org.apache.zookeeper.Op;
import org.apache.zookeeper.WatchedEvent;
import org.apache.zookeeper.Watcher;
import org.apache.zookeeper.data.Stat;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.slf4j.MDC;

import static org.apache.solr.common.cloud.ZkStateReader.BASE_URL_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.COLLECTION_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.CORE_NAME_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.CORE_NODE_NAME_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.ELECTION_NODE_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.NODE_NAME_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.REJOIN_AT_HEAD_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.SHARD_ID_PROP;

/**
 * Handle ZooKeeper interactions.
 * <p>
 * notes: loads everything on init, creates what's not there - further updates
 * are prompted with Watches.
 * <p>
 * TODO: exceptions during close on attempts to update cloud state
 */
public class ZkController {

    private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
    static final int WAIT_DOWN_STATES_TIMEOUT_SECONDS = 60;

    private final boolean SKIP_AUTO_RECOVERY = Boolean.getBoolean("solrcloud.skip.autorecovery");

    private final DistributedQueue overseerJobQueue;
    private final OverseerTaskQueue overseerCollectionQueue;
    private final OverseerTaskQueue overseerConfigSetQueue;

    private final DistributedMap overseerRunningMap;
    private final DistributedMap overseerCompletedMap;
    private final DistributedMap overseerFailureMap;

    public final static String COLLECTION_PARAM_PREFIX = "collection.";
    public final static String CONFIGNAME_PROP = "configName";

    static class ContextKey {

        private String collection;
        private String coreNodeName;

        public ContextKey(String collection, String coreNodeName) {
            this.collection = collection;
            this.coreNodeName = coreNodeName;
        }

        @Override
        public int hashCode() {
            final int prime = 31;
            int result = 1;
            result = prime * result + ((collection == null) ? 0 : collection.hashCode());
            result = prime * result + ((coreNodeName == null) ? 0 : coreNodeName.hashCode());
            return result;
        }

        @Override
        public boolean equals(Object obj) {
            if (this == obj)
                return true;
            if (obj == null)
                return false;
            if (getClass() != obj.getClass())
                return false;
            ContextKey other = (ContextKey) obj;
            if (collection == null) {
                if (other.collection != null)
                    return false;
            } else if (!collection.equals(other.collection))
                return false;
            if (coreNodeName == null) {
                if (other.coreNodeName != null)
                    return false;
            } else if (!coreNodeName.equals(other.coreNodeName))
                return false;
            return true;
        }
    }

    private final Map<ContextKey, ElectionContext> electionContexts = Collections.synchronizedMap(new HashMap<>());

    private final SolrZkClient zkClient;
    private final ZkCmdExecutor cmdExecutor;
    private final ZkStateReader zkStateReader;

    private final String zkServerAddress; // example: 127.0.0.1:54062/solr

    private final int localHostPort; // example: 54065
    private final String hostName; // example: 127.0.0.1
    private final String nodeName; // example: 127.0.0.1:54065_solr
    private String baseURL; // example: http://127.0.0.1:54065/solr

    private final CloudConfig cloudConfig;

    private LeaderElector overseerElector;

    // for now, this can be null in tests, in which case recovery will be inactive, and other features
    // may accept defaults or use mocks rather than pulling things from a CoreContainer
    private CoreContainer cc;

    protected volatile Overseer overseer;

    private int leaderVoteWait;
    private int leaderConflictResolveWait;

    private boolean genericCoreNodeNames;

    private int clientTimeout;

    private volatile boolean isClosed;

    // keeps track of replicas that have been asked to recover by leaders running on this node
    private final Map<String, String> replicasInLeaderInitiatedRecovery = new HashMap<String, String>();

    // This is an expert and unsupported development mode that does not create
    // an Overseer or register a /live node. This let's you monitor the cluster
    // and interact with zookeeper via the Solr admin UI on a node outside the cluster,
    // and so one that will not be killed or stopped when testing. See developer cloud-scripts.
    private boolean zkRunOnly = Boolean.getBoolean("zkRunOnly"); // expert

    // keeps track of a list of objects that need to know a new ZooKeeper session was created after expiration occurred
    // ref is held as a HashSet since we clone the set before notifying to avoid synchronizing too long
    private HashSet<OnReconnect> reconnectListeners = new HashSet<OnReconnect>();

    private class RegisterCoreAsync implements Callable {

        CoreDescriptor descriptor;
        boolean recoverReloadedCores;
        boolean afterExpiration;

        RegisterCoreAsync(CoreDescriptor descriptor, boolean recoverReloadedCores, boolean afterExpiration) {
            this.descriptor = descriptor;
            this.recoverReloadedCores = recoverReloadedCores;
            this.afterExpiration = afterExpiration;
        }

        public Object call() throws Exception {
            log.info("Registering core {} afterExpiration? {}", descriptor.getName(), afterExpiration);
            register(descriptor.getName(), descriptor, recoverReloadedCores, afterExpiration, false);
            return descriptor;
        }
    }

    // notifies registered listeners after the ZK reconnect in the background
    private class OnReconnectNotifyAsync implements Callable {

        private final OnReconnect listener;

        OnReconnectNotifyAsync(OnReconnect listener) {
            this.listener = listener;
        }

        @Override
        public Object call() throws Exception {
            listener.command();
            return null;
        }
    }

    public ZkController(final CoreContainer cc, String zkServerAddress, int zkClientConnectTimeout,
            CloudConfig cloudConfig, final CurrentCoreDescriptorProvider registerOnReconnect)
            throws InterruptedException, TimeoutException, IOException {

        if (cc == null)
            throw new IllegalArgumentException("CoreContainer cannot be null.");
        this.cc = cc;

        this.cloudConfig = cloudConfig;

        this.genericCoreNodeNames = cloudConfig.getGenericCoreNodeNames();

        // be forgiving and strip this off leading/trailing slashes
        // this allows us to support users specifying hostContext="/" in 
        // solr.xml to indicate the root context, instead of hostContext="" 
        // which means the default of "solr"
        String localHostContext = trimLeadingAndTrailingSlashes(cloudConfig.getSolrHostContext());

        this.zkServerAddress = zkServerAddress;
        this.localHostPort = cloudConfig.getSolrHostPort();
        this.hostName = normalizeHostName(cloudConfig.getHost());
        this.nodeName = generateNodeName(this.hostName, Integer.toString(this.localHostPort), localHostContext);
        MDCLoggingContext.setNode(nodeName);
        this.leaderVoteWait = cloudConfig.getLeaderVoteWait();
        this.leaderConflictResolveWait = cloudConfig.getLeaderConflictResolveWait();

        this.clientTimeout = cloudConfig.getZkClientTimeout();
        DefaultConnectionStrategy strat = new DefaultConnectionStrategy();
        String zkACLProviderClass = cloudConfig.getZkACLProviderClass();
        ZkACLProvider zkACLProvider = null;
        if (zkACLProviderClass != null && zkACLProviderClass.trim().length() > 0) {
            zkACLProvider = cc.getResourceLoader().newInstance(zkACLProviderClass, ZkACLProvider.class);
        } else {
            zkACLProvider = new DefaultZkACLProvider();
        }

        String zkCredentialsProviderClass = cloudConfig.getZkCredentialsProviderClass();
        if (zkCredentialsProviderClass != null && zkCredentialsProviderClass.trim().length() > 0) {
            strat.setZkCredentialsToAddAutomatically(
                    cc.getResourceLoader().newInstance(zkCredentialsProviderClass, ZkCredentialsProvider.class));
        } else {
            strat.setZkCredentialsToAddAutomatically(new DefaultZkCredentialsProvider());
        }
        addOnReconnectListener(getConfigDirListener());

        zkClient = new SolrZkClient(zkServerAddress, clientTimeout, zkClientConnectTimeout, strat,
                // on reconnect, reload cloud info
                new OnReconnect() {

                    @Override
                    public void command() {
                        log.info(
                                "ZooKeeper session re-connected ... refreshing core states after session expiration.");

                        try {
                            zkStateReader.createClusterStateWatchersAndUpdate();

                            // this is troublesome - we dont want to kill anything the old
                            // leader accepted
                            // though I guess sync will likely get those updates back? But
                            // only if
                            // he is involved in the sync, and he certainly may not be
                            // ExecutorUtil.shutdownAndAwaitTermination(cc.getCmdDistribExecutor());
                            // we need to create all of our lost watches

                            // seems we dont need to do this again...
                            // Overseer.createClientNodes(zkClient, getNodeName());

                            // start the overseer first as following code may need it's processing
                            if (!zkRunOnly) {
                                ElectionContext context = new OverseerElectionContext(zkClient, overseer,
                                        getNodeName());

                                ElectionContext prevContext = overseerElector.getContext();
                                if (prevContext != null) {
                                    prevContext.cancelElection();
                                }

                                overseerElector.setup(context);
                                overseerElector.joinElection(context, true);
                            }

                            cc.cancelCoreRecoveries();

                            registerAllCoresAsDown(registerOnReconnect, false);

                            // we have to register as live first to pick up docs in the buffer
                            createEphemeralLiveNode();

                            List<CoreDescriptor> descriptors = registerOnReconnect.getCurrentDescriptors();
                            // re register all descriptors
                            ExecutorService executorService = (cc != null) ? cc.getCoreZkRegisterExecutorService()
                                    : null;
                            if (descriptors != null) {
                                for (CoreDescriptor descriptor : descriptors) {
                                    // TODO: we need to think carefully about what happens when it
                                    // was
                                    // a leader that was expired - as well as what to do about
                                    // leaders/overseers
                                    // with connection loss
                                    try {
                                        // unload solrcores that have been 'failed over'
                                        throwErrorIfReplicaReplaced(descriptor);

                                        if (executorService != null) {
                                            executorService.submit(new RegisterCoreAsync(descriptor, true, true));
                                        } else {
                                            register(descriptor.getName(), descriptor, true, true, false);
                                        }
                                    } catch (Exception e) {
                                        SolrException.log(log, "Error registering SolrCore", e);
                                    }
                                }
                            }

                            // notify any other objects that need to know when the session was re-connected
                            HashSet<OnReconnect> clonedListeners;
                            synchronized (reconnectListeners) {
                                clonedListeners = (HashSet<OnReconnect>) reconnectListeners.clone();
                            }
                            // the OnReconnect operation can be expensive per listener, so do that async in the background
                            for (OnReconnect listener : clonedListeners) {
                                try {
                                    if (executorService != null) {
                                        executorService.submit(new OnReconnectNotifyAsync(listener));
                                    } else {
                                        listener.command();
                                    }
                                } catch (Exception exc) {
                                    // not much we can do here other than warn in the log
                                    log.warn("Error when notifying OnReconnect listener " + listener
                                            + " after session re-connected.", exc);
                                }
                            }
                        } catch (InterruptedException e) {
                            // Restore the interrupted status
                            Thread.currentThread().interrupt();
                            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
                        } catch (Exception e) {
                            SolrException.log(log, "", e);
                            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
                        }
                    }

                }, new BeforeReconnect() {

                    @Override
                    public void command() {
                        try {
                            ZkController.this.overseer.close();
                        } catch (Exception e) {
                            log.error("Error trying to stop any Overseer threads", e);
                        }
                        closeOutstandingElections(registerOnReconnect);
                        markAllAsNotLeader(registerOnReconnect);
                    }
                }, zkACLProvider);

        this.overseerJobQueue = Overseer.getStateUpdateQueue(zkClient);
        this.overseerCollectionQueue = Overseer.getCollectionQueue(zkClient);
        this.overseerConfigSetQueue = Overseer.getConfigSetQueue(zkClient);
        this.overseerRunningMap = Overseer.getRunningMap(zkClient);
        this.overseerCompletedMap = Overseer.getCompletedMap(zkClient);
        this.overseerFailureMap = Overseer.getFailureMap(zkClient);
        cmdExecutor = new ZkCmdExecutor(clientTimeout);
        zkStateReader = new ZkStateReader(zkClient, () -> {
            if (cc != null)
                cc.securityNodeChanged();
        });

        init(registerOnReconnect);
    }

    public int getLeaderVoteWait() {
        return leaderVoteWait;
    }

    public int getLeaderConflictResolveWait() {
        return leaderConflictResolveWait;
    }

    private void registerAllCoresAsDown(final CurrentCoreDescriptorProvider registerOnReconnect,
            boolean updateLastPublished) {
        List<CoreDescriptor> descriptors = registerOnReconnect.getCurrentDescriptors();
        if (isClosed)
            return;
        if (descriptors != null) {
            // before registering as live, make sure everyone is in a
            // down state
            publishNodeAsDown(getNodeName());
            for (CoreDescriptor descriptor : descriptors) {
                // if it looks like we are going to be the leader, we don't
                // want to wait for the following stuff
                CloudDescriptor cloudDesc = descriptor.getCloudDescriptor();
                String collection = cloudDesc.getCollectionName();
                String slice = cloudDesc.getShardId();
                try {

                    int children = zkStateReader.getZkClient().getChildren(ZkStateReader.COLLECTIONS_ZKNODE + "/"
                            + collection + "/leader_elect/" + slice + "/election", null, true).size();
                    if (children == 0) {
                        log.debug("looks like we are going to be the leader for collection {} shard {}", collection,
                                slice);
                        continue;
                    }

                } catch (NoNodeException e) {
                    log.debug("looks like we are going to be the leader for collection {} shard {}", collection,
                            slice);
                    continue;
                } catch (InterruptedException e2) {
                    Thread.currentThread().interrupt();
                } catch (KeeperException e) {
                    log.warn("", e);
                    Thread.currentThread().interrupt();
                }

                final String coreZkNodeName = descriptor.getCloudDescriptor().getCoreNodeName();
                try {
                    log.debug("calling waitForLeaderToSeeDownState for coreZkNodeName={} collection={} shard={}",
                            new Object[] { coreZkNodeName, collection, slice });
                    waitForLeaderToSeeDownState(descriptor, coreZkNodeName);
                } catch (Exception e) {
                    SolrException.log(log, "", e);
                    if (isClosed) {
                        return;
                    }
                }
            }
        }
    }

    private void closeOutstandingElections(final CurrentCoreDescriptorProvider registerOnReconnect) {

        List<CoreDescriptor> descriptors = registerOnReconnect.getCurrentDescriptors();
        if (descriptors != null) {
            for (CoreDescriptor descriptor : descriptors) {
                closeExistingElectionContext(descriptor);
            }
        }
    }

    private ContextKey closeExistingElectionContext(CoreDescriptor cd) {
        // look for old context - if we find it, cancel it
        String collection = cd.getCloudDescriptor().getCollectionName();
        final String coreNodeName = cd.getCloudDescriptor().getCoreNodeName();

        ContextKey contextKey = new ContextKey(collection, coreNodeName);
        ElectionContext prevContext = electionContexts.get(contextKey);

        if (prevContext != null) {
            prevContext.close();
            electionContexts.remove(contextKey);
        }

        return contextKey;
    }

    private void markAllAsNotLeader(final CurrentCoreDescriptorProvider registerOnReconnect) {
        List<CoreDescriptor> descriptors = registerOnReconnect.getCurrentDescriptors();
        if (descriptors != null) {
            for (CoreDescriptor descriptor : descriptors) {
                descriptor.getCloudDescriptor().setLeader(false);
                descriptor.getCloudDescriptor().setHasRegistered(false);
            }
        }
    }

    /**
     * Closes the underlying ZooKeeper client.
     */
    public void close() {
        this.isClosed = true;
        try {
            for (ElectionContext context : electionContexts.values()) {
                try {
                    context.close();
                } catch (Exception e) {
                    log.error("Error closing overseer", e);
                }
            }
        } finally {
            try {
                try {
                    overseer.close();
                } catch (Exception e) {
                    log.error("Error closing overseer", e);
                }
            } finally {
                try {
                    try {
                        zkStateReader.close();
                    } catch (Exception e) {
                        log.error("Error closing zkStateReader", e);
                    }
                } finally {
                    try {
                        zkClient.close();
                    } catch (Exception e) {
                        log.error("Error closing zkClient", e);
                    }
                }
            }
        }

    }

    /**
     * Returns true if config file exists
     */
    public boolean configFileExists(String collection, String fileName)
            throws KeeperException, InterruptedException {
        Stat stat = zkClient.exists(ZkConfigManager.CONFIGS_ZKNODE + "/" + collection + "/" + fileName, null, true);
        return stat != null;
    }

    /**
     * @return information about the cluster from ZooKeeper
     */
    public ClusterState getClusterState() {
        return zkStateReader.getClusterState();
    }

    /**
     * Returns config file data (in bytes)
     */
    public byte[] getConfigFileData(String zkConfigName, String fileName)
            throws KeeperException, InterruptedException {
        String zkPath = ZkConfigManager.CONFIGS_ZKNODE + "/" + zkConfigName + "/" + fileName;
        byte[] bytes = zkClient.getData(zkPath, null, null, true);
        if (bytes == null) {
            log.error("Config file contains no data:" + zkPath);
            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR,
                    "Config file contains no data:" + zkPath);
        }

        return bytes;
    }

    // normalize host removing any url scheme.
    // input can be null, host, or url_prefix://host
    private String normalizeHostName(String host) throws IOException {

        if (host == null || host.length() == 0) {
            String hostaddress;
            try {
                hostaddress = InetAddress.getLocalHost().getHostAddress();
            } catch (UnknownHostException e) {
                hostaddress = "127.0.0.1"; // cannot resolve system hostname, fall through
            }
            // Re-get the IP again for "127.0.0.1", the other case we trust the hosts
            // file is right.
            if ("127.0.0.1".equals(hostaddress)) {
                Enumeration<NetworkInterface> netInterfaces = null;
                try {
                    netInterfaces = NetworkInterface.getNetworkInterfaces();
                    while (netInterfaces.hasMoreElements()) {
                        NetworkInterface ni = netInterfaces.nextElement();
                        Enumeration<InetAddress> ips = ni.getInetAddresses();
                        while (ips.hasMoreElements()) {
                            InetAddress ip = ips.nextElement();
                            if (ip.isSiteLocalAddress()) {
                                hostaddress = ip.getHostAddress();
                            }
                        }
                    }
                } catch (Exception e) {
                    SolrException.log(log, "Error while looking for a better host name than 127.0.0.1", e);
                }
            }
            host = hostaddress;
        } else {
            if (URLUtil.hasScheme(host)) {
                host = URLUtil.removeScheme(host);
            }
        }

        return host;
    }

    public String getHostName() {
        return hostName;
    }

    public int getHostPort() {
        return localHostPort;
    }

    public SolrZkClient getZkClient() {
        return zkClient;
    }

    /**
     * @return zookeeper server address
     */
    public String getZkServerAddress() {
        return zkServerAddress;
    }

    /**
     * Create the zknodes necessary for a cluster to operate
     *
     * @param zkClient a SolrZkClient
     * @throws KeeperException      if there is a Zookeeper error
     * @throws InterruptedException on interrupt
     */
    public static void createClusterZkNodes(SolrZkClient zkClient) throws KeeperException, InterruptedException {
        ZkCmdExecutor cmdExecutor = new ZkCmdExecutor(zkClient.getZkClientTimeout());
        cmdExecutor.ensureExists(ZkStateReader.LIVE_NODES_ZKNODE, zkClient);
        cmdExecutor.ensureExists(ZkStateReader.COLLECTIONS_ZKNODE, zkClient);
        cmdExecutor.ensureExists(ZkStateReader.ALIASES, zkClient);
        byte[] emptyJson = "{}".getBytes(StandardCharsets.UTF_8);
        cmdExecutor.ensureExists(ZkStateReader.CLUSTER_STATE, emptyJson, CreateMode.PERSISTENT, zkClient);
        cmdExecutor.ensureExists(ZkStateReader.SOLR_SECURITY_CONF_PATH, emptyJson, CreateMode.PERSISTENT, zkClient);
    }

    private void init(CurrentCoreDescriptorProvider registerOnReconnect) {

        try {
            createClusterZkNodes(zkClient);
            zkStateReader.createClusterStateWatchersAndUpdate();
            this.baseURL = zkStateReader.getBaseUrlForNodeName(this.nodeName);

            checkForExistingEphemeralNode();

            // start the overseer first as following code may need it's processing
            if (!zkRunOnly) {
                overseerElector = new LeaderElector(zkClient);
                this.overseer = new Overseer(cc.getShardHandlerFactory().getShardHandler(),
                        cc.getUpdateShardHandler(), CommonParams.CORES_HANDLER_PATH, zkStateReader, this,
                        cloudConfig);
                ElectionContext context = new OverseerElectionContext(zkClient, overseer, getNodeName());
                overseerElector.setup(context);
                overseerElector.joinElection(context, false);
            }

            Stat stat = zkClient.exists(ZkStateReader.LIVE_NODES_ZKNODE, null, true);
            if (stat != null && stat.getNumChildren() > 0) {
                publishAndWaitForDownStates();
            }

            // Do this last to signal we're up.
            createEphemeralLiveNode();
        } catch (IOException e) {
            log.error("", e);
            throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Can't create ZooKeeperController", e);
        } catch (InterruptedException e) {
            // Restore the interrupted status
            Thread.currentThread().interrupt();
            log.error("", e);
            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
        } catch (KeeperException e) {
            log.error("", e);
            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
        }

    }

    private void checkForExistingEphemeralNode() throws KeeperException, InterruptedException {
        if (zkRunOnly) {
            return;
        }
        String nodeName = getNodeName();
        String nodePath = ZkStateReader.LIVE_NODES_ZKNODE + "/" + nodeName;

        if (!zkClient.exists(nodePath, true)) {
            return;
        }

        final CountDownLatch deletedLatch = new CountDownLatch(1);
        Stat stat = zkClient.exists(nodePath, event -> {
            if (Watcher.Event.EventType.None.equals(event.getType())) {
                return;
            }
            if (Watcher.Event.EventType.NodeDeleted.equals(event.getType())) {
                deletedLatch.countDown();
            }
        }, true);

        if (stat == null) {
            // znode suddenly disappeared but that's okay
            return;
        }

        boolean deleted = deletedLatch.await(zkClient.getSolrZooKeeper().getSessionTimeout() * 2,
                TimeUnit.MILLISECONDS);
        if (!deleted) {
            throw new SolrException(ErrorCode.SERVER_ERROR, "A previous ephemeral live node still exists. "
                    + "Solr cannot continue. Please ensure that no other Solr process using the same port is running already.");
        }
    }

    public void publishAndWaitForDownStates() throws KeeperException, InterruptedException {

        publishNodeAsDown(getNodeName());

        Set<String> collectionsWithLocalReplica = ConcurrentHashMap.newKeySet();
        for (SolrCore core : cc.getCores()) {
            collectionsWithLocalReplica.add(core.getCoreDescriptor().getCloudDescriptor().getCollectionName());
        }

        CountDownLatch latch = new CountDownLatch(collectionsWithLocalReplica.size());
        for (String collectionWithLocalReplica : collectionsWithLocalReplica) {
            zkStateReader.registerCollectionStateWatcher(collectionWithLocalReplica,
                    (liveNodes, collectionState) -> {
                        boolean foundStates = true;
                        for (SolrCore core : cc.getCores()) {
                            if (core.getCoreDescriptor().getCloudDescriptor().getCollectionName()
                                    .equals(collectionWithLocalReplica)) {
                                Replica replica = collectionState.getReplica(
                                        core.getCoreDescriptor().getCloudDescriptor().getCoreNodeName());
                                if (replica.getState() != Replica.State.DOWN) {
                                    foundStates = false;
                                }
                            }
                        }

                        if (foundStates && collectionsWithLocalReplica.remove(collectionWithLocalReplica)) {
                            latch.countDown();
                        }
                        return foundStates;
                    });
        }

        boolean allPublishedDown = latch.await(WAIT_DOWN_STATES_TIMEOUT_SECONDS, TimeUnit.SECONDS);
        if (!allPublishedDown) {
            log.warn("Timed out waiting to see all nodes published as DOWN in our cluster state.");
        }
    }

    /**
     * Validates if the chroot exists in zk (or if it is successfully created).
     * Optionally, if create is set to true this method will create the path in
     * case it doesn't exist
     *
     * @return true if the path exists or is created false if the path doesn't
     * exist and 'create' = false
     */
    public static boolean checkChrootPath(String zkHost, boolean create)
            throws KeeperException, InterruptedException {
        if (!SolrZkClient.containsChroot(zkHost)) {
            return true;
        }
        log.trace("zkHost includes chroot");
        String chrootPath = zkHost.substring(zkHost.indexOf("/"), zkHost.length());

        SolrZkClient tmpClient = new SolrZkClient(zkHost.substring(0, zkHost.indexOf("/")), 60000, 30000, null,
                null, null);
        boolean exists = tmpClient.exists(chrootPath, true);
        if (!exists && create) {
            tmpClient.makePath(chrootPath, false, true);
            exists = true;
        }
        tmpClient.close();
        return exists;
    }

    public boolean isConnected() {
        return zkClient.isConnected();
    }

    private void createEphemeralLiveNode() throws KeeperException, InterruptedException {
        if (zkRunOnly) {
            return;
        }
        String nodeName = getNodeName();
        String nodePath = ZkStateReader.LIVE_NODES_ZKNODE + "/" + nodeName;
        log.info("Register node as live in ZooKeeper:" + nodePath);
        zkClient.makePath(nodePath, CreateMode.EPHEMERAL, true);
    }

    public String getNodeName() {
        return nodeName;
    }

    /**
     * Returns true if the path exists
     */
    public boolean pathExists(String path) throws KeeperException, InterruptedException {
        return zkClient.exists(path, true);
    }

    /**
     * Register shard with ZooKeeper.
     *
     * @return the shardId for the SolrCore
     */
    public String register(String coreName, final CoreDescriptor desc, boolean skipRecovery) throws Exception {
        return register(coreName, desc, false, false, skipRecovery);
    }

    /**
     * Register shard with ZooKeeper.
     *
     * @return the shardId for the SolrCore
     */
    public String register(String coreName, final CoreDescriptor desc, boolean recoverReloadedCores,
            boolean afterExpiration, boolean skipRecovery) throws Exception {
        try (SolrCore core = cc.getCore(desc.getName())) {
            MDCLoggingContext.setCore(core);
        }
        try {
            // pre register has published our down state
            final String baseUrl = getBaseUrl();

            final CloudDescriptor cloudDesc = desc.getCloudDescriptor();
            final String collection = cloudDesc.getCollectionName();

            final String coreZkNodeName = desc.getCloudDescriptor().getCoreNodeName();
            assert coreZkNodeName != null : "we should have a coreNodeName by now";

            String shardId = cloudDesc.getShardId();
            Map<String, Object> props = new HashMap<>();
            // we only put a subset of props into the leader node
            props.put(ZkStateReader.BASE_URL_PROP, baseUrl);
            props.put(ZkStateReader.CORE_NAME_PROP, coreName);
            props.put(ZkStateReader.NODE_NAME_PROP, getNodeName());

            log.debug("Register replica - core:{} address:{} collection:{} shard:{}", coreName, baseUrl,
                    cloudDesc.getCollectionName(), shardId);

            ZkNodeProps leaderProps = new ZkNodeProps(props);

            try {
                // If we're a preferred leader, insert ourselves at the head of the queue
                boolean joinAtHead = false;
                Replica replica = zkStateReader.getClusterState()
                        .getReplica(desc.getCloudDescriptor().getCollectionName(), coreZkNodeName);
                if (replica != null) {
                    joinAtHead = replica.getBool(SliceMutator.PREFERRED_LEADER_PROP, false);
                }
                joinElection(desc, afterExpiration, joinAtHead);
            } catch (InterruptedException e) {
                // Restore the interrupted status
                Thread.currentThread().interrupt();
                throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
            } catch (KeeperException | IOException e) {
                throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
            }

            // in this case, we want to wait for the leader as long as the leader might
            // wait for a vote, at least - but also long enough that a large cluster has
            // time to get its act together
            String leaderUrl = getLeader(cloudDesc, leaderVoteWait + 600000);

            String ourUrl = ZkCoreNodeProps.getCoreUrl(baseUrl, coreName);
            log.debug("We are " + ourUrl + " and leader is " + leaderUrl);
            boolean isLeader = leaderUrl.equals(ourUrl);

            try (SolrCore core = cc.getCore(desc.getName())) {

                // recover from local transaction log and wait for it to complete before
                // going active
                // TODO: should this be moved to another thread? To recoveryStrat?
                // TODO: should this actually be done earlier, before (or as part of)
                // leader election perhaps?

                UpdateLog ulog = core.getUpdateHandler().getUpdateLog();

                // we will call register again after zk expiration and on reload
                if (!afterExpiration && !core.isReloaded() && ulog != null) {
                    // disable recovery in case shard is in construction state (for shard splits)
                    Slice slice = getClusterState().getSlice(collection, shardId);
                    if (slice.getState() != Slice.State.CONSTRUCTION || !isLeader) {
                        Future<UpdateLog.RecoveryInfo> recoveryFuture = core.getUpdateHandler().getUpdateLog()
                                .recoverFromLog();
                        if (recoveryFuture != null) {
                            log.info("Replaying tlog for " + ourUrl
                                    + " during startup... NOTE: This can take a while.");
                            recoveryFuture.get(); // NOTE: this could potentially block for
                            // minutes or more!
                            // TODO: public as recovering in the mean time?
                            // TODO: in the future we could do peersync in parallel with recoverFromLog
                        } else {
                            log.debug("No LogReplay needed for core={} baseURL={}", core.getName(), baseUrl);
                        }
                    }
                }
                boolean didRecovery = checkRecovery(recoverReloadedCores, isLeader, skipRecovery, collection,
                        coreZkNodeName, core, cc, afterExpiration);
                if (!didRecovery) {
                    publish(desc, Replica.State.ACTIVE);
                }

                core.getCoreDescriptor().getCloudDescriptor().setHasRegistered(true);
            }

            // make sure we have an update cluster state right away
            zkStateReader.forceUpdateCollection(collection);
            return shardId;
        } finally {
            MDCLoggingContext.clear();
        }
    }

    // timeoutms is the timeout for the first call to get the leader - there is then
    // a longer wait to make sure that leader matches our local state
    private String getLeader(final CloudDescriptor cloudDesc, int timeoutms) {

        String collection = cloudDesc.getCollectionName();
        String shardId = cloudDesc.getShardId();
        // rather than look in the cluster state file, we go straight to the zknodes
        // here, because on cluster restart there could be stale leader info in the
        // cluster state node that won't be updated for a moment
        String leaderUrl;
        try {
            leaderUrl = getLeaderProps(collection, cloudDesc.getShardId(), timeoutms).getCoreUrl();

            // now wait until our currently cloud state contains the latest leader
            String clusterStateLeaderUrl = zkStateReader.getLeaderUrl(collection, shardId, timeoutms * 2); // since we found it in zk, we are willing to
            // wait a while to find it in state
            int tries = 0;
            final long msInSec = 1000L;
            int maxTries = (int) Math.floor(leaderConflictResolveWait / msInSec);
            while (!leaderUrl.equals(clusterStateLeaderUrl)) {
                if (tries > maxTries) {
                    throw new SolrException(ErrorCode.SERVER_ERROR,
                            "There is conflicting information about the leader of shard: " + cloudDesc.getShardId()
                                    + " our state says:" + clusterStateLeaderUrl + " but zookeeper says:"
                                    + leaderUrl);
                }
                tries++;
                if (tries % 30 == 0) {
                    String warnMsg = String.format(Locale.ENGLISH,
                            "Still seeing conflicting information about the leader "
                                    + "of shard %s for collection %s after %d seconds; our state says %s, but ZooKeeper says %s",
                            cloudDesc.getShardId(), collection, tries, clusterStateLeaderUrl, leaderUrl);
                    log.warn(warnMsg);
                }
                Thread.sleep(msInSec);
                clusterStateLeaderUrl = zkStateReader.getLeaderUrl(collection, shardId, timeoutms);
                leaderUrl = getLeaderProps(collection, cloudDesc.getShardId(), timeoutms).getCoreUrl();
            }

        } catch (Exception e) {
            log.error("Error getting leader from zk", e);
            throw new SolrException(SolrException.ErrorCode.SERVER_ERROR,
                    "Error getting leader from zk for shard " + shardId, e);
        }
        return leaderUrl;
    }

    /**
     * Get leader props directly from zk nodes.
     */
    public ZkCoreNodeProps getLeaderProps(final String collection, final String slice, int timeoutms)
            throws InterruptedException {
        return getLeaderProps(collection, slice, timeoutms, false);
    }

    /**
     * Get leader props directly from zk nodes.
     *
     * @return leader props
     */
    public ZkCoreNodeProps getLeaderProps(final String collection, final String slice, int timeoutms,
            boolean failImmediatelyOnExpiration) throws InterruptedException {
        int iterCount = timeoutms / 1000;
        Exception exp = null;
        while (iterCount-- > 0) {
            try {
                byte[] data = zkClient.getData(ZkStateReader.getShardLeadersPath(collection, slice), null, null,
                        true);
                ZkCoreNodeProps leaderProps = new ZkCoreNodeProps(ZkNodeProps.load(data));
                return leaderProps;
            } catch (InterruptedException e) {
                throw e;
            } catch (SessionExpiredException e) {
                if (failImmediatelyOnExpiration) {
                    throw new RuntimeException("Session has expired - could not get leader props", exp);
                }
                exp = e;
                Thread.sleep(1000);
            } catch (Exception e) {
                exp = e;
                Thread.sleep(1000);
            }
            if (cc.isShutDown()) {
                throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE, "CoreContainer is closed");
            }
        }
        throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE, "Could not get leader props", exp);
    }

    private void joinElection(CoreDescriptor cd, boolean afterExpiration, boolean joinAtHead)
            throws InterruptedException, KeeperException, IOException {
        // look for old context - if we find it, cancel it
        String collection = cd.getCloudDescriptor().getCollectionName();
        final String coreNodeName = cd.getCloudDescriptor().getCoreNodeName();

        ContextKey contextKey = new ContextKey(collection, coreNodeName);

        ElectionContext prevContext = electionContexts.get(contextKey);

        if (prevContext != null) {
            prevContext.cancelElection();
        }

        String shardId = cd.getCloudDescriptor().getShardId();

        Map<String, Object> props = new HashMap<>();
        // we only put a subset of props into the leader node
        props.put(ZkStateReader.BASE_URL_PROP, getBaseUrl());
        props.put(ZkStateReader.CORE_NAME_PROP, cd.getName());
        props.put(ZkStateReader.NODE_NAME_PROP, getNodeName());
        props.put(ZkStateReader.CORE_NODE_NAME_PROP, coreNodeName);

        ZkNodeProps ourProps = new ZkNodeProps(props);

        LeaderElector leaderElector = new LeaderElector(zkClient, contextKey, electionContexts);
        ElectionContext context = new ShardLeaderElectionContext(leaderElector, shardId, collection, coreNodeName,
                ourProps, this, cc);

        leaderElector.setup(context);
        electionContexts.put(contextKey, context);
        leaderElector.joinElection(context, false, joinAtHead);
    }

    /**
     * Returns whether or not a recovery was started
     */
    private boolean checkRecovery(boolean recoverReloadedCores, final boolean isLeader, boolean skipRecovery,
            final String collection, String shardId, SolrCore core, CoreContainer cc, boolean afterExpiration) {
        if (SKIP_AUTO_RECOVERY) {
            log.warn("Skipping recovery according to sys prop solrcloud.skip.autorecovery");
            return false;
        }
        boolean doRecovery = true;
        if (!isLeader) {

            if (skipRecovery || (!afterExpiration && core.isReloaded() && !recoverReloadedCores)) {
                doRecovery = false;
            }

            if (doRecovery) {
                log.info("Core needs to recover:" + core.getName());
                core.getUpdateHandler().getSolrCoreState().doRecovery(cc, core.getCoreDescriptor());
                return true;
            }

            // see if the leader told us to recover
            final Replica.State lirState = getLeaderInitiatedRecoveryState(collection, shardId,
                    core.getCoreDescriptor().getCloudDescriptor().getCoreNodeName());
            if (lirState == Replica.State.DOWN) {
                log.info("Leader marked core " + core.getName() + " down; starting recovery process");
                core.getUpdateHandler().getSolrCoreState().doRecovery(cc, core.getCoreDescriptor());
                return true;
            }
        } else {
            log.info("I am the leader, no recovery necessary");
        }

        return false;
    }

    public String getBaseUrl() {
        return baseURL;
    }

    public void publish(final CoreDescriptor cd, final Replica.State state)
            throws KeeperException, InterruptedException {
        publish(cd, state, true);
    }

    public void publish(final CoreDescriptor cd, final Replica.State state, boolean updateLastState)
            throws KeeperException, InterruptedException {
        publish(cd, state, updateLastState, false);
    }

    /**
     * Publish core state to overseer.
     */
    public void publish(final CoreDescriptor cd, final Replica.State state, boolean updateLastState,
            boolean forcePublish) throws KeeperException, InterruptedException {
        if (!forcePublish) {
            try (SolrCore core = cc.getCore(cd.getName())) {
                if (core == null || core.isClosed()) {
                    return;
                }
                MDCLoggingContext.setCore(core);
            }
        } else {
            MDCLoggingContext.setCoreDescriptor(cd);
        }
        try {
            String collection = cd.getCloudDescriptor().getCollectionName();

            log.debug("publishing state={}", state.toString());
            // System.out.println(Thread.currentThread().getStackTrace()[3]);
            Integer numShards = cd.getCloudDescriptor().getNumShards();
            if (numShards == null) { // XXX sys prop hack
                log.debug("numShards not found on descriptor - reading it from system property");
                numShards = Integer.getInteger(ZkStateReader.NUM_SHARDS_PROP);
            }

            assert collection != null && collection.length() > 0;

            String shardId = cd.getCloudDescriptor().getShardId();

            String coreNodeName = cd.getCloudDescriptor().getCoreNodeName();
            // If the leader initiated recovery, then verify that this replica has performed
            // recovery as requested before becoming active; don't even look at lirState if going down
            if (state != Replica.State.DOWN) {
                final Replica.State lirState = getLeaderInitiatedRecoveryState(collection, shardId, coreNodeName);
                if (lirState != null) {
                    if (state == Replica.State.ACTIVE) {
                        // trying to become active, so leader-initiated state must be recovering
                        if (lirState == Replica.State.RECOVERING) {
                            updateLeaderInitiatedRecoveryState(collection, shardId, coreNodeName,
                                    Replica.State.ACTIVE, cd, true);
                        } else if (lirState == Replica.State.DOWN) {
                            throw new SolrException(ErrorCode.INVALID_STATE, "Cannot publish state of core '"
                                    + cd.getName() + "' as active without recovering first!");
                        }
                    } else if (state == Replica.State.RECOVERING) {
                        // if it is currently DOWN, then trying to enter into recovering state is good
                        if (lirState == Replica.State.DOWN) {
                            updateLeaderInitiatedRecoveryState(collection, shardId, coreNodeName,
                                    Replica.State.RECOVERING, cd, true);
                        }
                    }
                }
            }

            Map<String, Object> props = new HashMap<>();
            props.put(Overseer.QUEUE_OPERATION, "state");
            props.put(ZkStateReader.STATE_PROP, state.toString());
            props.put(ZkStateReader.BASE_URL_PROP, getBaseUrl());
            props.put(ZkStateReader.CORE_NAME_PROP, cd.getName());
            props.put(ZkStateReader.ROLES_PROP, cd.getCloudDescriptor().getRoles());
            props.put(ZkStateReader.NODE_NAME_PROP, getNodeName());
            props.put(ZkStateReader.SHARD_ID_PROP, cd.getCloudDescriptor().getShardId());
            props.put(ZkStateReader.COLLECTION_PROP, collection);
            if (numShards != null) {
                props.put(ZkStateReader.NUM_SHARDS_PROP, numShards.toString());
            }
            if (coreNodeName != null) {
                props.put(ZkStateReader.CORE_NODE_NAME_PROP, coreNodeName);
            }
            try (SolrCore core = cc.getCore(cd.getName())) {
                if (core != null && core.getDirectoryFactory().isSharedStorage()) {
                    if (core != null && core.getDirectoryFactory().isSharedStorage()) {
                        props.put("dataDir", core.getDataDir());
                        UpdateLog ulog = core.getUpdateHandler().getUpdateLog();
                        if (ulog != null) {
                            props.put("ulogDir", ulog.getLogDir());
                        }
                    }
                }
            }

            ZkNodeProps m = new ZkNodeProps(props);

            if (updateLastState) {
                cd.getCloudDescriptor().lastPublished = state;
            }
            overseerJobQueue.offer(Utils.toJSON(m));
        } finally {
            MDCLoggingContext.clear();
        }
    }

    private boolean needsToBeAssignedShardId(final CoreDescriptor desc, final ClusterState state,
            final String coreNodeName) {

        final CloudDescriptor cloudDesc = desc.getCloudDescriptor();

        final String shardId = state.getShardId(getNodeName(), desc.getName());

        if (shardId != null) {
            cloudDesc.setShardId(shardId);
            return false;
        }
        return true;
    }

    public void unregister(String coreName, CoreDescriptor cd) throws InterruptedException, KeeperException {
        final String coreNodeName = cd.getCloudDescriptor().getCoreNodeName();
        final String collection = cd.getCloudDescriptor().getCollectionName();

        if (Strings.isNullOrEmpty(collection)) {
            log.error("No collection was specified.");
            assert false : "No collection was specified [" + collection + "]";
            return;
        }

        ElectionContext context = electionContexts.remove(new ContextKey(collection, coreNodeName));

        if (context != null) {
            context.cancelElection();
        }

        CloudDescriptor cloudDescriptor = cd.getCloudDescriptor();
        zkStateReader.unregisterCore(cloudDescriptor.getCollectionName());

        ZkNodeProps m = new ZkNodeProps(Overseer.QUEUE_OPERATION, OverseerAction.DELETECORE.toLower(),
                ZkStateReader.CORE_NAME_PROP, coreName, ZkStateReader.NODE_NAME_PROP, getNodeName(),
                ZkStateReader.COLLECTION_PROP, cloudDescriptor.getCollectionName(),
                ZkStateReader.CORE_NODE_NAME_PROP, coreNodeName);
        overseerJobQueue.offer(Utils.toJSON(m));
    }

    public void createCollection(String collection) throws KeeperException, InterruptedException {
        ZkNodeProps m = new ZkNodeProps(Overseer.QUEUE_OPERATION,
                CollectionParams.CollectionAction.CREATE.toLower(), ZkStateReader.NODE_NAME_PROP, getNodeName(),
                ZkStateReader.COLLECTION_PROP, collection);
        overseerJobQueue.offer(Utils.toJSON(m));
    }

    // convenience for testing
    void printLayoutToStdOut() throws KeeperException, InterruptedException {
        zkClient.printLayoutToStdOut();
    }

    public ZkStateReader getZkStateReader() {
        return zkStateReader;
    }

    private void doGetShardIdAndNodeNameProcess(CoreDescriptor cd) {
        final String coreNodeName = cd.getCloudDescriptor().getCoreNodeName();

        if (coreNodeName != null) {
            waitForShardId(cd);
        } else {
            // if no explicit coreNodeName, we want to match by base url and core name
            waitForCoreNodeName(cd);
            waitForShardId(cd);
        }
    }

    private void waitForCoreNodeName(CoreDescriptor descriptor) {
        int retryCount = 320;
        log.debug("look for our core node name");
        while (retryCount-- > 0) {
            Map<String, Slice> slicesMap = zkStateReader.getClusterState()
                    .getSlicesMap(descriptor.getCloudDescriptor().getCollectionName());
            if (slicesMap != null) {

                for (Slice slice : slicesMap.values()) {
                    for (Replica replica : slice.getReplicas()) {
                        // TODO: for really large clusters, we could 'index' on this

                        String nodeName = replica.getStr(ZkStateReader.NODE_NAME_PROP);
                        String core = replica.getStr(ZkStateReader.CORE_NAME_PROP);

                        String msgNodeName = getNodeName();
                        String msgCore = descriptor.getName();

                        if (msgNodeName.equals(nodeName) && core.equals(msgCore)) {
                            descriptor.getCloudDescriptor().setCoreNodeName(replica.getName());
                            return;
                        }
                    }
                }
            }
            try {
                Thread.sleep(1000);
            } catch (InterruptedException e) {
                Thread.currentThread().interrupt();
            }
        }
    }

    private void waitForShardId(CoreDescriptor cd) {
        log.debug("waiting to find shard id in clusterstate for " + cd.getName());
        int retryCount = 320;
        while (retryCount-- > 0) {
            final String shardId = zkStateReader.getClusterState().getShardId(cd.getCollectionName(), getNodeName(),
                    cd.getName());
            if (shardId != null) {
                cd.getCloudDescriptor().setShardId(shardId);
                return;
            }
            try {
                Thread.sleep(1000);
            } catch (InterruptedException e) {
                Thread.currentThread().interrupt();
            }
        }

        throw new SolrException(ErrorCode.SERVER_ERROR, "Could not get shard id for core: " + cd.getName());
    }

    public String getCoreNodeName(CoreDescriptor descriptor) {
        String coreNodeName = descriptor.getCloudDescriptor().getCoreNodeName();
        if (coreNodeName == null && !genericCoreNodeNames) {
            // it's the default
            return getNodeName() + "_" + descriptor.getName();
        }

        return coreNodeName;
    }

    public void preRegister(CoreDescriptor cd) {

        String coreNodeName = getCoreNodeName(cd);

        // before becoming available, make sure we are not live and active
        // this also gets us our assigned shard id if it was not specified
        try {
            checkStateInZk(cd);

            CloudDescriptor cloudDesc = cd.getCloudDescriptor();

            // make sure the node name is set on the descriptor
            if (cloudDesc.getCoreNodeName() == null) {
                cloudDesc.setCoreNodeName(coreNodeName);
            }

            publish(cd, Replica.State.DOWN, false, true);
            String collectionName = cd.getCloudDescriptor().getCollectionName();
            DocCollection collection = zkStateReader.getClusterState().getCollectionOrNull(collectionName);
            log.debug(collection == null
                    ? "Collection {} not visible yet, but flagging it so a watch is registered when it becomes visible"
                    : "Registering watch for collection {}", collectionName);
            zkStateReader.registerCore(collectionName);
        } catch (KeeperException e) {
            log.error("", e);
            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
        } catch (InterruptedException e) {
            Thread.currentThread().interrupt();
            log.error("", e);
            throw new ZooKeeperException(SolrException.ErrorCode.SERVER_ERROR, "", e);
        }

        if (cd.getCloudDescriptor().getShardId() == null
                && needsToBeAssignedShardId(cd, zkStateReader.getClusterState(), coreNodeName)) {
            doGetShardIdAndNodeNameProcess(cd);
        } else {
            // still wait till we see us in local state
            doGetShardIdAndNodeNameProcess(cd);
        }

    }

    private void checkStateInZk(CoreDescriptor cd) throws InterruptedException {
        if (!Overseer.isLegacy(zkStateReader)) {
            CloudDescriptor cloudDesc = cd.getCloudDescriptor();
            String coreNodeName = cloudDesc.getCoreNodeName();
            if (coreNodeName == null)
                throw new SolrException(ErrorCode.SERVER_ERROR, "No coreNodeName for " + cd);
            if (cloudDesc.getShardId() == null) {
                throw new SolrException(ErrorCode.SERVER_ERROR, "No shard id for " + cd);
            }

            AtomicReference<String> errorMessage = new AtomicReference<>();
            AtomicReference<DocCollection> collectionState = new AtomicReference<>();
            try {
                zkStateReader.waitForState(cd.getCollectionName(), 3, TimeUnit.SECONDS, (n, c) -> {
                    collectionState.set(c);
                    if (c == null)
                        return false;
                    Slice slice = c.getSlice(cloudDesc.getShardId());
                    if (slice == null) {
                        errorMessage.set("Invalid shard: " + cloudDesc.getShardId());
                        return false;
                    }
                    Replica replica = slice.getReplica(coreNodeName);
                    if (replica == null) {
                        errorMessage.set("coreNodeName " + coreNodeName + " does not exist in shard "
                                + cloudDesc.getShardId());
                        return false;
                    }
                    String baseUrl = replica.getStr(BASE_URL_PROP);
                    String coreName = replica.getStr(CORE_NAME_PROP);
                    if (baseUrl.equals(this.baseURL) && coreName.equals(cd.getName())) {
                        return true;
                    }
                    errorMessage.set("coreNodeName " + coreNodeName
                            + " exists, but does not match expected node or core name");
                    return false;
                });
            } catch (TimeoutException e) {
                String error = errorMessage.get();
                if (error == null)
                    error = "Replica " + coreNodeName + " is not present in cluster state";
                throw new SolrException(ErrorCode.SERVER_ERROR, error + ": " + collectionState.get());
            }
        }
    }

    private ZkCoreNodeProps waitForLeaderToSeeDownState(CoreDescriptor descriptor, final String coreZkNodeName) {
        // try not to wait too long here - if we are waiting too long, we should probably
        // move along and join the election

        CloudDescriptor cloudDesc = descriptor.getCloudDescriptor();
        String collection = cloudDesc.getCollectionName();
        String shard = cloudDesc.getShardId();
        ZkCoreNodeProps leaderProps = null;

        int retries = 2;
        for (int i = 0; i < retries; i++) {
            try {
                if (isClosed) {
                    throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE, "We have been closed");
                }

                // go straight to zk, not the cloud state - we want current info
                leaderProps = getLeaderProps(collection, shard, 5000);
                break;
            } catch (Exception e) {
                SolrException.log(log, "There was a problem finding the leader in zk", e);
                try {
                    Thread.sleep(2000);
                } catch (InterruptedException e1) {
                    Thread.currentThread().interrupt();
                }
                if (i == retries - 1) {
                    throw new SolrException(ErrorCode.SERVER_ERROR, "There was a problem finding the leader in zk");
                }
            }
        }

        String leaderBaseUrl = leaderProps.getBaseUrl();
        String leaderCoreName = leaderProps.getCoreName();

        String myCoreNodeName = cloudDesc.getCoreNodeName();
        String myCoreName = descriptor.getName();
        String ourUrl = ZkCoreNodeProps.getCoreUrl(getBaseUrl(), myCoreName);

        boolean isLeader = leaderProps.getCoreUrl().equals(ourUrl);
        if (!isLeader && !SKIP_AUTO_RECOVERY) {

            // detect if this core is in leader-initiated recovery and if so, 
            // then we don't need the leader to wait on seeing the down state
            Replica.State lirState = null;
            try {
                lirState = getLeaderInitiatedRecoveryState(collection, shard, myCoreNodeName);
            } catch (Exception exc) {
                log.error("Failed to determine if replica " + myCoreNodeName
                        + " is in leader-initiated recovery due to: " + exc, exc);
            }

            if (lirState != null) {
                log.debug("Replica " + myCoreNodeName
                        + " is already in leader-initiated recovery, so not waiting for leader to see down state.");
            } else {

                log.info("Replica " + myCoreNodeName
                        + " NOT in leader-initiated recovery, need to wait for leader to see down state.");

                try (HttpSolrClient client = new Builder(leaderBaseUrl).build()) {
                    client.setConnectionTimeout(15000);
                    client.setSoTimeout(120000);
                    WaitForState prepCmd = new WaitForState();
                    prepCmd.setCoreName(leaderCoreName);
                    prepCmd.setNodeName(getNodeName());
                    prepCmd.setCoreNodeName(coreZkNodeName);
                    prepCmd.setState(Replica.State.DOWN);

                    // let's retry a couple times - perhaps the leader just went down,
                    // or perhaps he is just not quite ready for us yet
                    retries = 2;
                    for (int i = 0; i < retries; i++) {
                        if (isClosed) {
                            throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE, "We have been closed");
                        }
                        try {
                            client.request(prepCmd);
                            break;
                        } catch (Exception e) {

                            // if the core container is shutdown, don't wait
                            if (cc.isShutDown()) {
                                throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE,
                                        "Core container is shutdown.");
                            }

                            Throwable rootCause = SolrException.getRootCause(e);
                            if (rootCause instanceof IOException) {
                                // if there was a communication error talking to the leader, see if the leader is even alive
                                if (!zkStateReader.getClusterState().liveNodesContain(leaderProps.getNodeName())) {
                                    throw new SolrException(ErrorCode.SERVICE_UNAVAILABLE,
                                            "Node " + leaderProps.getNodeName() + " hosting leader for " + shard
                                                    + " in " + collection + " is not live!");
                                }
                            }

                            SolrException.log(log, "There was a problem making a request to the leader", e);
                            try {
                                Thread.sleep(2000);
                            } catch (InterruptedException e1) {
                                Thread.currentThread().interrupt();
                            }
                            if (i == retries - 1) {
                                throw new SolrException(ErrorCode.SERVER_ERROR,
                                        "There was a problem making a request to the leader");
                            }
                        }
                    }
                } catch (IOException e) {
                    SolrException.log(log, "Error closing HttpSolrClient", e);
                }
            }
        }
        return leaderProps;
    }

    public static void linkConfSet(SolrZkClient zkClient, String collection, String confSetName)
            throws KeeperException, InterruptedException {
        String path = ZkStateReader.COLLECTIONS_ZKNODE + "/" + collection;
        log.debug("Load collection config from:" + path);
        byte[] data;
        try {
            data = zkClient.getData(path, null, null, true);
        } catch (NoNodeException e) {
            // if there is no node, we will try and create it
            // first try to make in case we are pre configuring
            ZkNodeProps props = new ZkNodeProps(CONFIGNAME_PROP, confSetName);
            try {

                zkClient.makePath(path, Utils.toJSON(props), CreateMode.PERSISTENT, null, true);
            } catch (KeeperException e2) {
                // it's okay if the node already exists
                if (e2.code() != KeeperException.Code.NODEEXISTS) {
                    throw e;
                }
                // if we fail creating, setdata
                // TODO: we should consider using version
                zkClient.setData(path, Utils.toJSON(props), true);
            }
            return;
        }
        // we found existing data, let's update it
        ZkNodeProps props = null;
        if (data != null) {
            props = ZkNodeProps.load(data);
            Map<String, Object> newProps = new HashMap<>();
            newProps.putAll(props.getProperties());
            newProps.put(CONFIGNAME_PROP, confSetName);
            props = new ZkNodeProps(newProps);
        } else {
            props = new ZkNodeProps(CONFIGNAME_PROP, confSetName);
        }

        // TODO: we should consider using version
        zkClient.setData(path, Utils.toJSON(props), true);

    }

    /**
     * If in SolrCloud mode, upload config sets for each SolrCore in solr.xml.
     */
    public static void bootstrapConf(SolrZkClient zkClient, CoreContainer cc, String solrHome) throws IOException {

        ZkConfigManager configManager = new ZkConfigManager(zkClient);

        //List<String> allCoreNames = cfg.getAllCoreNames();
        List<CoreDescriptor> cds = cc.getCoresLocator().discover(cc);

        log.info(
                "bootstrapping config for " + cds.size() + " cores into ZooKeeper using solr.xml from " + solrHome);

        for (CoreDescriptor cd : cds) {
            String coreName = cd.getName();
            String confName = cd.getCollectionName();
            if (StringUtils.isEmpty(confName))
                confName = coreName;
            Path udir = cd.getInstanceDir().resolve("conf");
            log.info("Uploading directory " + udir + " with name " + confName + " for SolrCore " + coreName);
            configManager.uploadConfigDir(udir, confName);
        }
    }

    public DistributedQueue getOverseerJobQueue() {
        return overseerJobQueue;
    }

    public OverseerTaskQueue getOverseerCollectionQueue() {
        return overseerCollectionQueue;
    }

    public OverseerTaskQueue getOverseerConfigSetQueue() {
        return overseerConfigSetQueue;
    }

    public DistributedMap getOverseerRunningMap() {
        return overseerRunningMap;
    }

    public DistributedMap getOverseerCompletedMap() {
        return overseerCompletedMap;
    }

    public DistributedMap getOverseerFailureMap() {
        return overseerFailureMap;
    }

    public int getClientTimeout() {
        return clientTimeout;
    }

    public Overseer getOverseer() {
        return overseer;
    }

    public LeaderElector getOverseerElector() {
        return overseerElector;
    }

    /**
     * Returns the nodeName that should be used based on the specified properties.
     *
     * @param hostName    - must not be null or the empty string
     * @param hostPort    - must consist only of digits, must not be null or the empty string
     * @param hostContext - should not begin or end with a slash (leading/trailin slashes will be ignored), must not be null, may be the empty string to denote the root context
     * @lucene.experimental
     * @see ZkStateReader#getBaseUrlForNodeName
     */
    static String generateNodeName(final String hostName, final String hostPort, final String hostContext) {
        try {
            return hostName + ':' + hostPort + '_'
                    + URLEncoder.encode(trimLeadingAndTrailingSlashes(hostContext), "UTF-8");
        } catch (UnsupportedEncodingException e) {
            throw new Error("JVM Does not seem to support UTF-8", e);
        }
    }

    /**
     * Utility method for trimming and leading and/or trailing slashes from
     * its input.  May return the empty string.  May return null if and only
     * if the input is null.
     */
    public static String trimLeadingAndTrailingSlashes(final String in) {
        if (null == in)
            return in;

        String out = in;
        if (out.startsWith("/")) {
            out = out.substring(1);
        }
        if (out.endsWith("/")) {
            out = out.substring(0, out.length() - 1);
        }
        return out;
    }

    public void rejoinOverseerElection(String electionNode, boolean joinAtHead) {
        try {
            if (electionNode != null) {
                //this call is from inside the JVM  . not from CoreAdminHandler
                if (overseerElector.getContext() == null || overseerElector.getContext().leaderSeqPath == null) {
                    overseerElector.retryElection(new OverseerElectionContext(zkClient, overseer, getNodeName()),
                            joinAtHead);
                    return;
                }
                if (!overseerElector.getContext().leaderSeqPath.endsWith(electionNode)) {
                    log.warn("Asked to rejoin with wrong election node : {}, current node is {}", electionNode,
                            overseerElector.getContext().leaderSeqPath);
                    //however delete it . This is possible when the last attempt at deleting the election node failed.
                    if (electionNode.startsWith(getNodeName())) {
                        try {
                            zkClient.delete(OverseerElectionContext.OVERSEER_ELECT + LeaderElector.ELECTION_NODE
                                    + "/" + electionNode, -1, true);
                        } catch (NoNodeException e) {
                            //no problem
                        } catch (InterruptedException e) {
                            Thread.currentThread().interrupt();
                        } catch (Exception e) {
                            log.warn("Old election node exists , could not be removed ", e);
                        }
                    }
                }
            } else {
                overseerElector.retryElection(overseerElector.getContext(), joinAtHead);
            }
        } catch (Exception e) {
            throw new SolrException(ErrorCode.SERVER_ERROR, "Unable to rejoin election", e);
        }

    }

    public void rejoinShardLeaderElection(SolrParams params) {
        try {

            String collectionName = params.get(COLLECTION_PROP);
            String shardId = params.get(SHARD_ID_PROP);
            String coreNodeName = params.get(CORE_NODE_NAME_PROP);
            String coreName = params.get(CORE_NAME_PROP);
            String electionNode = params.get(ELECTION_NODE_PROP);
            String baseUrl = params.get(BASE_URL_PROP);

            try (SolrCore core = cc.getCore(coreName)) {
                MDCLoggingContext.setCore(core);

                log.info("Rejoin the shard leader election.");

                ContextKey contextKey = new ContextKey(collectionName, coreNodeName);

                ElectionContext prevContext = electionContexts.get(contextKey);
                if (prevContext != null)
                    prevContext.cancelElection();

                ZkNodeProps zkProps = new ZkNodeProps(BASE_URL_PROP, baseUrl, CORE_NAME_PROP, coreName,
                        NODE_NAME_PROP, getNodeName(), CORE_NODE_NAME_PROP, coreNodeName);

                LeaderElector elect = ((ShardLeaderElectionContextBase) prevContext).getLeaderElector();
                ShardLeaderElectionContext context = new ShardLeaderElectionContext(elect, shardId, collectionName,
                        coreNodeName, zkProps, this, getCoreContainer());

                context.leaderSeqPath = context.electionPath + LeaderElector.ELECTION_NODE + "/" + electionNode;
                elect.setup(context);
                electionContexts.put(contextKey, context);

                elect.retryElection(context, params.getBool(REJOIN_AT_HEAD_PROP, false));
            }
        } catch (Exception e) {
            throw new SolrException(ErrorCode.SERVER_ERROR, "Unable to rejoin election", e);
        }

    }

    public void checkOverseerDesignate() {
        try {
            byte[] data = zkClient.getData(ZkStateReader.ROLES, null, new Stat(), true);
            if (data == null)
                return;
            Map roles = (Map) Utils.fromJSON(data);
            if (roles == null)
                return;
            List nodeList = (List) roles.get("overseer");
            if (nodeList == null)
                return;
            if (nodeList.contains(getNodeName())) {
                ZkNodeProps props = new ZkNodeProps(Overseer.QUEUE_OPERATION,
                        CollectionParams.CollectionAction.ADDROLE.toString().toLowerCase(Locale.ROOT), "node",
                        getNodeName(), "role", "overseer");
                log.info("Going to add role {} ", props);
                getOverseerCollectionQueue().offer(Utils.toJSON(props));
            }
        } catch (NoNodeException nne) {
            return;
        } catch (Exception e) {
            log.warn("could not read the overseer designate ", e);
        }
    }

    CoreContainer getCoreContainer() {
        return cc;
    }

    /**
     * When a leader receives a communication error when trying to send a request to a replica,
     * it calls this method to ensure the replica enters recovery when connectivity is restored.
     * <p>
     * returns true if the node hosting the replica is still considered "live" by ZooKeeper;
     * false means the node is not live either, so no point in trying to send recovery commands
     * to it.
     */
    public boolean ensureReplicaInLeaderInitiatedRecovery(final CoreContainer container, final String collection,
            final String shardId, final ZkCoreNodeProps replicaCoreProps, CoreDescriptor leaderCd,
            boolean forcePublishState) throws KeeperException, InterruptedException {
        final String replicaUrl = replicaCoreProps.getCoreUrl();

        if (collection == null)
            throw new IllegalArgumentException(
                    "collection parameter cannot be null for starting leader-initiated recovery for replica: "
                            + replicaUrl);

        if (shardId == null)
            throw new IllegalArgumentException(
                    "shard parameter cannot be null for starting leader-initiated recovery for replica: "
                            + replicaUrl);

        if (replicaUrl == null)
            throw new IllegalArgumentException(
                    "replicaUrl parameter cannot be null for starting leader-initiated recovery");

        // First, determine if this replica is already in recovery handling
        // which is needed because there can be many concurrent errors flooding in
        // about the same replica having trouble and we only need to send the "needs"
        // recovery signal once
        boolean nodeIsLive = true;
        String replicaNodeName = replicaCoreProps.getNodeName();
        String replicaCoreNodeName = ((Replica) replicaCoreProps.getNodeProps()).getName();
        assert replicaCoreNodeName != null : "No core name for replica " + replicaNodeName;
        synchronized (replicasInLeaderInitiatedRecovery) {
            if (replicasInLeaderInitiatedRecovery.containsKey(replicaUrl)) {
                if (!forcePublishState) {
                    log.debug("Replica {} already in leader-initiated recovery handling.", replicaUrl);
                    return false; // already in this recovery process
                }
            }

            // we only really need to try to start the LIR process if the node itself is "live"
            if (getZkStateReader().getClusterState().liveNodesContain(replicaNodeName)) {

                LeaderInitiatedRecoveryThread lirThread = new LeaderInitiatedRecoveryThread(this, container,
                        collection, shardId, replicaCoreProps, 120, leaderCd);
                ExecutorService executor = container.getUpdateShardHandler().getUpdateExecutor();
                try {
                    MDC.put("DistributedUpdateProcessor.replicaUrlToRecover", replicaCoreProps.getCoreUrl());
                    executor.execute(lirThread);
                } finally {
                    MDC.remove("DistributedUpdateProcessor.replicaUrlToRecover");
                }

                // create a znode that requires the replica needs to "ack" to verify it knows it was out-of-sync
                replicasInLeaderInitiatedRecovery.put(replicaUrl,
                        getLeaderInitiatedRecoveryZnodePath(collection, shardId, replicaCoreNodeName));
                log.info(
                        "Put replica core={} coreNodeName={} on " + replicaNodeName
                                + " into leader-initiated recovery.",
                        replicaCoreProps.getCoreName(), replicaCoreNodeName);
            } else {
                nodeIsLive = false; // we really don't need to send the recovery request if the node is NOT live
                log.info("Node " + replicaNodeName
                        + " is not live, so skipping leader-initiated recovery for replica: core={} coreNodeName={}",
                        replicaCoreProps.getCoreName(), replicaCoreNodeName);
                // publishDownState will be false to avoid publishing the "down" state too many times
                // as many errors can occur together and will each call into this method (SOLR-6189)        
            }
        }

        return nodeIsLive;
    }

    public boolean isReplicaInRecoveryHandling(String replicaUrl) {
        boolean exists = false;
        synchronized (replicasInLeaderInitiatedRecovery) {
            exists = replicasInLeaderInitiatedRecovery.containsKey(replicaUrl);
        }
        return exists;
    }

    public void removeReplicaFromLeaderInitiatedRecoveryHandling(String replicaUrl) {
        synchronized (replicasInLeaderInitiatedRecovery) {
            replicasInLeaderInitiatedRecovery.remove(replicaUrl);
        }
    }

    public Replica.State getLeaderInitiatedRecoveryState(String collection, String shardId, String coreNodeName) {
        final Map<String, Object> stateObj = getLeaderInitiatedRecoveryStateObject(collection, shardId,
                coreNodeName);
        if (stateObj == null) {
            return null;
        }
        final String stateStr = (String) stateObj.get(ZkStateReader.STATE_PROP);
        return stateStr == null ? null : Replica.State.getState(stateStr);
    }

    public Map<String, Object> getLeaderInitiatedRecoveryStateObject(String collection, String shardId,
            String coreNodeName) {

        if (collection == null || shardId == null || coreNodeName == null)
            return null; // if we don't have complete data about a core in cloud mode, return null

        String znodePath = getLeaderInitiatedRecoveryZnodePath(collection, shardId, coreNodeName);
        byte[] stateData = null;
        try {
            stateData = zkClient.getData(znodePath, null, new Stat(), false);
        } catch (NoNodeException ignoreMe) {
            // safe to ignore as this znode will only exist if the leader initiated recovery
        } catch (ConnectionLossException | SessionExpiredException cle) {
            // sort of safe to ignore ??? Usually these are seen when the core is going down
            // or there are bigger issues to deal with than reading this znode
            log.warn("Unable to read " + znodePath + " due to: " + cle);
        } catch (Exception exc) {
            log.error("Failed to read data from znode " + znodePath + " due to: " + exc);
            if (exc instanceof SolrException) {
                throw (SolrException) exc;
            } else {
                throw new SolrException(ErrorCode.SERVER_ERROR, "Failed to read data from znodePath: " + znodePath,
                        exc);
            }
        }

        Map<String, Object> stateObj = null;
        if (stateData != null && stateData.length > 0) {
            // TODO: Remove later ... this is for upgrading from 4.8.x to 4.10.3 (see: SOLR-6732)
            if (stateData[0] == (byte) '{') {
                Object parsedJson = Utils.fromJSON(stateData);
                if (parsedJson instanceof Map) {
                    stateObj = (Map<String, Object>) parsedJson;
                } else {
                    throw new SolrException(ErrorCode.SERVER_ERROR,
                            "Leader-initiated recovery state data is invalid! " + parsedJson);
                }
            } else {
                // old format still in ZK
                stateObj = Utils.makeMap("state", new String(stateData, StandardCharsets.UTF_8));
            }
        }

        return stateObj;
    }

    public void updateLeaderInitiatedRecoveryState(String collection, String shardId, String coreNodeName,
            Replica.State state, CoreDescriptor leaderCd, boolean retryOnConnLoss) {
        if (collection == null || shardId == null || coreNodeName == null) {
            log.warn("Cannot set leader-initiated recovery state znode to " + state.toString()
                    + " using: collection=" + collection + "; shardId=" + shardId + "; coreNodeName="
                    + coreNodeName);
            return; // if we don't have complete data about a core in cloud mode, do nothing
        }

        assert leaderCd != null;
        assert leaderCd.getCloudDescriptor() != null;

        String leaderCoreNodeName = leaderCd.getCloudDescriptor().getCoreNodeName();

        String znodePath = getLeaderInitiatedRecoveryZnodePath(collection, shardId, coreNodeName);

        if (state == Replica.State.ACTIVE) {
            // since we're marking it active, we don't need this znode anymore, so delete instead of update
            try {
                zkClient.delete(znodePath, -1, retryOnConnLoss);
            } catch (Exception justLogIt) {
                log.warn("Failed to delete znode " + znodePath, justLogIt);
            }
            return;
        }

        Map<String, Object> stateObj = null;
        try {
            stateObj = getLeaderInitiatedRecoveryStateObject(collection, shardId, coreNodeName);
        } catch (Exception exc) {
            log.warn(exc.getMessage(), exc);
        }
        if (stateObj == null) {
            stateObj = Utils.makeMap();
        }

        stateObj.put(ZkStateReader.STATE_PROP, state.toString());
        // only update the createdBy value if it's not set
        if (stateObj.get("createdByNodeName") == null) {
            stateObj.put("createdByNodeName", this.nodeName);
        }
        if (stateObj.get("createdByCoreNodeName") == null && leaderCoreNodeName != null) {
            stateObj.put("createdByCoreNodeName", leaderCoreNodeName);
        }

        byte[] znodeData = Utils.toJSON(stateObj);

        try {
            if (state == Replica.State.DOWN) {
                markShardAsDownIfLeader(collection, shardId, leaderCd, znodePath, znodeData, retryOnConnLoss);
            } else {
                // must retry on conn loss otherwise future election attempts may assume wrong LIR state
                if (zkClient.exists(znodePath, true)) {
                    zkClient.setData(znodePath, znodeData, retryOnConnLoss);
                } else {
                    zkClient.makePath(znodePath, znodeData, retryOnConnLoss);
                }
            }
            log.debug("Wrote {} to {}", state.toString(), znodePath);
        } catch (Exception exc) {
            if (exc instanceof SolrException) {
                throw (SolrException) exc;
            } else {
                throw new SolrException(ErrorCode.SERVER_ERROR,
                        "Failed to update data to " + state.toString() + " for znode: " + znodePath, exc);
            }
        }
    }

    /**
     * we use ZK's multi-transactional semantics to ensure that we are able to
     * publish a replica as 'down' only if our leader election node still exists
     * in ZK. This ensures that a long running network partition caused by GC etc
     * doesn't let us mark a node as down *after* we've already lost our session
     */
    private void markShardAsDownIfLeader(String collection, String shardId, CoreDescriptor leaderCd,
            String znodePath, byte[] znodeData, boolean retryOnConnLoss)
            throws KeeperException, InterruptedException {

        if (!leaderCd.getCloudDescriptor().isLeader()) {
            log.info("No longer leader, aborting attempt to mark shard down as part of LIR");
            throw new NotLeaderException(ErrorCode.SERVER_ERROR, "Locally, we do not think we are the leader.");
        }

        ContextKey key = new ContextKey(collection, leaderCd.getCloudDescriptor().getCoreNodeName());
        ElectionContext context = electionContexts.get(key);

        // we make sure we locally think we are the leader before and after getting the context - then
        // we only try zk if we still think we are the leader and have our leader context
        if (context == null || !leaderCd.getCloudDescriptor().isLeader()) {
            log.info("No longer leader, aborting attempt to mark shard down as part of LIR");
            throw new NotLeaderException(ErrorCode.SERVER_ERROR, "Locally, we do not think we are the leader.");
        }

        // we think we are the leader - get the expected shard leader version
        // we use this version and multi to ensure *only* the current zk registered leader
        // for a shard can put a replica into LIR

        Integer leaderZkNodeParentVersion = ((ShardLeaderElectionContextBase) context)
                .getLeaderZkNodeParentVersion();

        // TODO: should we do this optimistically to avoid races?
        if (zkClient.exists(znodePath, retryOnConnLoss)) {
            List<Op> ops = new ArrayList<>(2);
            ops.add(Op.check(new org.apache.hadoop.fs.Path(((ShardLeaderElectionContextBase) context).leaderPath)
                    .getParent().toString(), leaderZkNodeParentVersion));
            ops.add(Op.setData(znodePath, znodeData, -1));
            zkClient.multi(ops, retryOnConnLoss);
        } else {
            String parentZNodePath = getLeaderInitiatedRecoveryZnodePath(collection, shardId);
            try {
                // make sure we don't create /collections/{collection} if they do not exist with 2 param
                zkClient.makePath(parentZNodePath, (byte[]) null, CreateMode.PERSISTENT, (Watcher) null, true,
                        retryOnConnLoss, 2);
            } catch (KeeperException.NodeExistsException nee) {
                // if it exists, that's great!
            }

            // we only create the entry if the context we are using is registered as the current leader in ZK
            List<Op> ops = new ArrayList<>(2);
            ops.add(Op.check(new org.apache.hadoop.fs.Path(((ShardLeaderElectionContextBase) context).leaderPath)
                    .getParent().toString(), leaderZkNodeParentVersion));
            ops.add(Op.create(znodePath, znodeData, zkClient.getZkACLProvider().getACLsToAdd(znodePath),
                    CreateMode.PERSISTENT));
            zkClient.multi(ops, retryOnConnLoss);
        }
    }

    public String getLeaderInitiatedRecoveryZnodePath(String collection, String shardId) {
        return "/collections/" + collection + "/leader_initiated_recovery/" + shardId;
    }

    public String getLeaderInitiatedRecoveryZnodePath(String collection, String shardId, String coreNodeName) {
        return getLeaderInitiatedRecoveryZnodePath(collection, shardId) + "/" + coreNodeName;
    }

    public void throwErrorIfReplicaReplaced(CoreDescriptor desc) {
        ClusterState clusterState = getZkStateReader().getClusterState();
        if (clusterState != null) {
            DocCollection collection = clusterState
                    .getCollectionOrNull(desc.getCloudDescriptor().getCollectionName());
            if (collection != null) {
                boolean autoAddReplicas = ClusterStateUtil.isAutoAddReplicas(getZkStateReader(),
                        collection.getName());
                if (autoAddReplicas) {
                    CloudUtil.checkSharedFSFailoverReplaced(cc, desc);
                }
            }
        }
    }

    /**
     * Add a listener to be notified once there is a new session created after a ZooKeeper session expiration occurs;
     * in most cases, listeners will be components that have watchers that need to be re-created.
     */
    public void addOnReconnectListener(OnReconnect listener) {
        if (listener != null) {
            synchronized (reconnectListeners) {
                reconnectListeners.add(listener);
                log.debug("Added new OnReconnect listener " + listener);
            }
        }
    }

    /**
     * Removed a previously registered OnReconnect listener, such as when a core is removed or reloaded.
     */
    public void removeOnReconnectListener(OnReconnect listener) {
        if (listener != null) {
            boolean wasRemoved;
            synchronized (reconnectListeners) {
                wasRemoved = reconnectListeners.remove(listener);
            }
            if (wasRemoved) {
                log.debug("Removed OnReconnect listener " + listener);
            } else {
                log.warn("Was asked to remove OnReconnect listener " + listener
                        + ", but remove operation did not find it in the list of registered listeners.");
            }
        }
    }

    Set<OnReconnect> getCurrentOnReconnectListeners() {
        HashSet<OnReconnect> clonedListeners;
        synchronized (reconnectListeners) {
            clonedListeners = (HashSet<OnReconnect>) reconnectListeners.clone();
        }
        return clonedListeners;
    }

    /**
     * Persists a config file to ZooKeeper using optimistic concurrency.
     *
     * @return true on success
     */
    public static int persistConfigResourceToZooKeeper(ZkSolrResourceLoader zkLoader, int znodeVersion,
            String resourceName, byte[] content, boolean createIfNotExists) {
        int latestVersion = znodeVersion;
        final ZkController zkController = zkLoader.getZkController();
        final SolrZkClient zkClient = zkController.getZkClient();
        final String resourceLocation = zkLoader.getConfigSetZkPath() + "/" + resourceName;
        String errMsg = "Failed to persist resource at {0} - old {1}";
        try {
            try {
                Stat stat = zkClient.setData(resourceLocation, content, znodeVersion, true);
                latestVersion = stat.getVersion();// if the set succeeded , it should have incremented the version by one always
                log.info("Persisted config data to node {} ", resourceLocation);
                touchConfDir(zkLoader);
            } catch (NoNodeException e) {
                if (createIfNotExists) {
                    try {
                        zkClient.create(resourceLocation, content, CreateMode.PERSISTENT, true);
                        latestVersion = 0;//just created so version must be zero
                        touchConfDir(zkLoader);
                    } catch (KeeperException.NodeExistsException nee) {
                        try {
                            Stat stat = zkClient.exists(resourceLocation, null, true);
                            log.debug("failed to set data version in zk is {} and expected version is {} ",
                                    stat.getVersion(), znodeVersion);
                        } catch (Exception e1) {
                            log.warn("could not get stat");
                        }

                        log.info(StrUtils.formatString(errMsg, resourceLocation, znodeVersion));
                        throw new ResourceModifiedInZkException(ErrorCode.CONFLICT,
                                StrUtils.formatString(errMsg, resourceLocation, znodeVersion) + ", retry.");
                    }
                }
            }

        } catch (KeeperException.BadVersionException bve) {
            int v = -1;
            try {
                Stat stat = zkClient.exists(resourceLocation, null, true);
                v = stat.getVersion();
            } catch (Exception e) {
                log.error(e.getMessage());

            }
            log.info(StrUtils.formatString(errMsg + " zkVersion= " + v, resourceLocation, znodeVersion));
            throw new ResourceModifiedInZkException(ErrorCode.CONFLICT,
                    StrUtils.formatString(errMsg, resourceLocation, znodeVersion) + ", retry.");
        } catch (ResourceModifiedInZkException e) {
            throw e;
        } catch (Exception e) {
            if (e instanceof InterruptedException) {
                Thread.currentThread().interrupt(); // Restore the interrupted status
            }
            final String msg = "Error persisting resource at " + resourceLocation;
            log.error(msg, e);
            throw new SolrException(ErrorCode.SERVER_ERROR, msg, e);
        }
        return latestVersion;
    }

    public static void touchConfDir(ZkSolrResourceLoader zkLoader) {
        SolrZkClient zkClient = zkLoader.getZkController().getZkClient();
        try {
            zkClient.setData(zkLoader.getConfigSetZkPath(), new byte[] { 0 }, true);
        } catch (Exception e) {
            if (e instanceof InterruptedException) {
                Thread.currentThread().interrupt(); // Restore the interrupted status
            }
            final String msg = "Error 'touching' conf location " + zkLoader.getConfigSetZkPath();
            log.error(msg, e);
            throw new SolrException(ErrorCode.SERVER_ERROR, msg, e);

        }
    }

    public static class ResourceModifiedInZkException extends SolrException {
        public ResourceModifiedInZkException(ErrorCode code, String msg) {
            super(code, msg);
        }
    }

    private void unregisterConfListener(String confDir, Runnable listener) {
        synchronized (confDirectoryListeners) {
            final Set<Runnable> listeners = confDirectoryListeners.get(confDir);
            if (listeners == null) {
                log.warn(confDir + " has no more registered listeners, but a live one attempted to unregister!");
                return;
            }
            if (listeners.remove(listener)) {
                log.debug("removed listener for config directory [{}]", confDir);
            }
            if (listeners.isEmpty()) {
                // no more listeners for this confDir, remove it from the map
                log.debug("No more listeners for config directory [{}]", confDir);
                confDirectoryListeners.remove(confDir);
            }
        }
    }

    /**
     * This will give a callback to the listener whenever a child is modified in the
     * conf directory. It is the responsibility of the listener to check if the individual
     * item of interest has been modified.  When the last core which was interested in
     * this conf directory is gone the listeners will be removed automatically.
     */
    public void registerConfListenerForCore(final String confDir, SolrCore core, final Runnable listener) {
        if (listener == null) {
            throw new NullPointerException("listener cannot be null");
        }
        synchronized (confDirectoryListeners) {
            final Set<Runnable> confDirListeners = getConfDirListeners(confDir);
            confDirListeners.add(listener);
            core.addCloseHook(new CloseHook() {
                @Override
                public void preClose(SolrCore core) {
                    unregisterConfListener(confDir, listener);
                }

                @Override
                public void postClose(SolrCore core) {
                }
            });
        }
    }

    // this method is called in a protected confDirListeners block
    private Set<Runnable> getConfDirListeners(final String confDir) {
        assert Thread.holdsLock(confDirectoryListeners) : "confDirListeners lock not held by thread";
        Set<Runnable> confDirListeners = confDirectoryListeners.get(confDir);
        if (confDirListeners == null) {
            log.debug("watch zkdir {}", confDir);
            confDirListeners = new HashSet<>();
            confDirectoryListeners.put(confDir, confDirListeners);
            setConfWatcher(confDir, new WatcherImpl(confDir), null);
        }
        return confDirListeners;
    }

    private final Map<String, Set<Runnable>> confDirectoryListeners = new HashMap<>();

    private class WatcherImpl implements Watcher {
        private final String zkDir;

        private WatcherImpl(String dir) {
            this.zkDir = dir;
        }

        @Override
        public void process(WatchedEvent event) {
            // session events are not change events, and do not remove the watcher
            if (Event.EventType.None.equals(event.getType())) {
                return;
            }

            Stat stat = null;
            try {
                stat = zkClient.exists(zkDir, null, true);
            } catch (KeeperException e) {
                //ignore , it is not a big deal
            } catch (InterruptedException e) {
                Thread.currentThread().interrupt();
            }

            boolean resetWatcher = false;
            try {
                resetWatcher = fireEventListeners(zkDir);
            } finally {
                if (Event.EventType.None.equals(event.getType())) {
                    log.debug("A node got unwatched for {}", zkDir);
                } else {
                    if (resetWatcher)
                        setConfWatcher(zkDir, this, stat);
                    else
                        log.debug("A node got unwatched for {}", zkDir);
                }
            }
        }
    }

    private boolean fireEventListeners(String zkDir) {
        synchronized (confDirectoryListeners) {
            // if this is not among directories to be watched then don't set the watcher anymore
            if (!confDirectoryListeners.containsKey(zkDir)) {
                log.debug("Watcher on {} is removed ", zkDir);
                return false;
            }
            final Set<Runnable> listeners = confDirectoryListeners.get(zkDir);
            if (listeners != null && !listeners.isEmpty()) {
                final Set<Runnable> listenersCopy = new HashSet<>(listeners);
                // run these in a separate thread because this can be long running
                new Thread(() -> {
                    log.debug("Running listeners for {}", zkDir);
                    for (final Runnable listener : listenersCopy) {
                        try {
                            listener.run();
                        } catch (Exception e) {
                            log.warn("listener throws error", e);
                        }
                    }
                }).start();

            }
        }
        return true;
    }

    private void setConfWatcher(String zkDir, Watcher watcher, Stat stat) {
        try {
            Stat newStat = zkClient.exists(zkDir, watcher, true);
            if (stat != null && newStat.getVersion() > stat.getVersion()) {
                //a race condition where a we missed an event fired
                //so fire the event listeners
                fireEventListeners(zkDir);
            }
        } catch (KeeperException e) {
            log.error("failed to set watcher for conf dir {} ", zkDir);
        } catch (InterruptedException e) {
            Thread.currentThread().interrupt();
            log.error("failed to set watcher for conf dir {} ", zkDir);
        }
    }

    public OnReconnect getConfigDirListener() {
        return () -> {
            synchronized (confDirectoryListeners) {
                for (String s : confDirectoryListeners.keySet()) {
                    setConfWatcher(s, new WatcherImpl(s), null);
                    fireEventListeners(s);
                }
            }
        };
    }

    public String getLeaderSeqPath(String collection, String coreNodeName) {
        ContextKey key = new ContextKey(collection, coreNodeName);
        ElectionContext context = electionContexts.get(key);
        return context != null ? context.leaderSeqPath : null;
    }

    /**
     * Thrown during leader initiated recovery process if current node is not leader
     */
    public static class NotLeaderException extends SolrException {
        public NotLeaderException(ErrorCode code, String msg) {
            super(code, msg);
        }
    }

    public boolean checkIfCoreNodeNameAlreadyExists(CoreDescriptor dcore) {
        DocCollection collection = zkStateReader.getClusterState().getCollectionOrNull(dcore.getCollectionName());
        if (collection != null) {
            Collection<Slice> slices = collection.getSlices();

            for (Slice slice : slices) {
                Collection<Replica> replicas = slice.getReplicas();
                for (Replica replica : replicas) {
                    if (replica.getName().equals(dcore.getCloudDescriptor().getCoreNodeName())) {
                        return true;
                    }
                }
            }
        }
        return false;
    }

    /**
     * Best effort to set DOWN state for all replicas on node.
     * 
     * @param nodeName to operate on
     */
    public void publishNodeAsDown(String nodeName) {
        log.debug("Publish node={} as DOWN", nodeName);
        ZkNodeProps m = new ZkNodeProps(Overseer.QUEUE_OPERATION, OverseerAction.DOWNNODE.toLower(),
                ZkStateReader.NODE_NAME_PROP, nodeName);
        try {
            Overseer.getStateUpdateQueue(getZkClient()).offer(Utils.toJSON(m));
        } catch (InterruptedException e) {
            Thread.interrupted();
            log.debug("Publish node as down was interrupted.");
        } catch (Exception e) {
            log.warn("Could not publish node as down: " + e.getMessage());
        }
    }
}