spark ApplicationInfo 源码

  • 2022-10-20
  • 浏览 (194)

spark ApplicationInfo 代码

文件路径:/core/src/main/scala/org/apache/spark/deploy/master/ApplicationInfo.scala

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *    http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.spark.deploy.master

import java.util.Date

import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer

import org.apache.spark.deploy.ApplicationDescription
import org.apache.spark.resource.{ResourceInformation, ResourceProfile, ResourceUtils}
import org.apache.spark.resource.ResourceProfile.DEFAULT_RESOURCE_PROFILE_ID
import org.apache.spark.rpc.RpcEndpointRef
import org.apache.spark.util.Utils

private[spark] class ApplicationInfo(
    val startTime: Long,
    val id: String,
    val desc: ApplicationDescription,
    val submitDate: Date,
    val driver: RpcEndpointRef,
    defaultCores: Int)
  extends Serializable {

  @transient var state: ApplicationState.Value = _
  @transient var executors: mutable.HashMap[Int, ExecutorDesc] = _
  @transient var removedExecutors: ArrayBuffer[ExecutorDesc] = _
  @transient var coresGranted: Int = _
  @transient var endTime: Long = _
  @transient var appSource: ApplicationSource = _

  @transient private var executorsPerResourceProfileId: mutable.HashMap[Int, mutable.Set[Int]] = _
  @transient private var targetNumExecutorsPerResourceProfileId: mutable.HashMap[Int, Int] = _
  @transient private var rpIdToResourceProfile: mutable.HashMap[Int, ResourceProfile] = _
  @transient private var rpIdToResourceDesc: mutable.HashMap[Int, ExecutorResourceDescription] = _

  @transient private var nextExecutorId: Int = _

  init()

  private def readObject(in: java.io.ObjectInputStream): Unit = Utils.tryOrIOException {
    in.defaultReadObject()
    init()
  }

  private def init(): Unit = {
    state = ApplicationState.WAITING
    executors = new mutable.HashMap[Int, ExecutorDesc]
    coresGranted = 0
    endTime = -1L
    appSource = new ApplicationSource(this)
    nextExecutorId = 0
    removedExecutors = new ArrayBuffer[ExecutorDesc]
    val initialExecutorLimit = desc.initialExecutorLimit.getOrElse(Integer.MAX_VALUE)

    rpIdToResourceProfile = new mutable.HashMap[Int, ResourceProfile]()
    rpIdToResourceProfile(DEFAULT_RESOURCE_PROFILE_ID) = desc.defaultProfile
    rpIdToResourceDesc = new mutable.HashMap[Int, ExecutorResourceDescription]()
    createResourceDescForResourceProfile(desc.defaultProfile)

    targetNumExecutorsPerResourceProfileId = new mutable.HashMap[Int, Int]()
    targetNumExecutorsPerResourceProfileId(DEFAULT_RESOURCE_PROFILE_ID) = initialExecutorLimit

    executorsPerResourceProfileId = new mutable.HashMap[Int, mutable.Set[Int]]()
  }

  private[deploy] def getOrUpdateExecutorsForRPId(rpId: Int): mutable.Set[Int] = {
    executorsPerResourceProfileId.getOrElseUpdate(rpId, mutable.HashSet[Int]())
  }

  private[deploy] def getTargetExecutorNumForRPId(rpId: Int): Int = {
    targetNumExecutorsPerResourceProfileId.getOrElse(rpId, 0)
  }

  private[deploy] def getRequestedRPIds(): Seq[Int] = {
    rpIdToResourceProfile.keys.toSeq.sorted
  }

  private def createResourceDescForResourceProfile(resourceProfile: ResourceProfile): Unit = {
    if (!rpIdToResourceDesc.contains(resourceProfile.id)) {
      val defaultMemoryMbPerExecutor = desc.memoryPerExecutorMB
      val defaultCoresPerExecutor = desc.coresPerExecutor
      val coresPerExecutor = resourceProfile.getExecutorCores
        .orElse(defaultCoresPerExecutor)
      val memoryMbPerExecutor = resourceProfile.getExecutorMemory
        .map(_.toInt)
        .getOrElse(defaultMemoryMbPerExecutor)
      val customResources = ResourceUtils.executorResourceRequestToRequirement(
        resourceProfile.getCustomExecutorResources().values.toSeq.sortBy(_.resourceName))

      rpIdToResourceDesc(resourceProfile.id) =
        ExecutorResourceDescription(coresPerExecutor, memoryMbPerExecutor, customResources)
    }
  }

  // Get resources required for schedule.
  private[deploy] def getResourceDescriptionForRpId(rpId: Int): ExecutorResourceDescription = {
    rpIdToResourceDesc(rpId)
  }

  private[deploy] def requestExecutors(
      resourceProfileToTotalExecs: Map[ResourceProfile, Int]): Unit = {
    resourceProfileToTotalExecs.foreach { case (rp, num) =>
      createResourceDescForResourceProfile(rp)

      if (!rpIdToResourceProfile.contains(rp.id)) {
        rpIdToResourceProfile(rp.id) = rp
      }

      targetNumExecutorsPerResourceProfileId(rp.id) = num
    }
  }

  private[deploy] def getResourceProfileById(rpId: Int): ResourceProfile = {
    rpIdToResourceProfile(rpId)
  }

  private def newExecutorId(useID: Option[Int] = None): Int = {
    useID match {
      case Some(id) =>
        nextExecutorId = math.max(nextExecutorId, id + 1)
        id
      case None =>
        val id = nextExecutorId
        nextExecutorId += 1
        id
    }
  }

  private[master] def addExecutor(
      worker: WorkerInfo,
      cores: Int,
      memoryMb: Int,
      resources: Map[String, ResourceInformation],
      rpId: Int,
      useID: Option[Int] = None): ExecutorDesc = {
    val exec = new ExecutorDesc(
      newExecutorId(useID), this, worker, cores, memoryMb, resources, rpId)
    executors(exec.id) = exec
    getOrUpdateExecutorsForRPId(rpId).add(exec.id)
    coresGranted += cores
    exec
  }

  private[master] def removeExecutor(exec: ExecutorDesc): Unit = {
    if (executors.contains(exec.id)) {
      removedExecutors += executors(exec.id)
      executors -= exec.id
      executorsPerResourceProfileId(exec.rpId) -= exec.id
      coresGranted -= exec.cores
    }
  }

  private val requestedCores = desc.maxCores.getOrElse(defaultCores)

  private[master] def coresLeft: Int = requestedCores - coresGranted

  private var _retryCount = 0

  private[master] def retryCount = _retryCount

  private[master] def incrementRetryCount() = {
    _retryCount += 1
    _retryCount
  }

  private[master] def resetRetryCount() = _retryCount = 0

  private[master] def markFinished(endState: ApplicationState.Value): Unit = {
    state = endState
    endTime = System.currentTimeMillis()
  }

  private[master] def isFinished: Boolean = {
    state != ApplicationState.WAITING && state != ApplicationState.RUNNING
  }

  /**
   * Return the total limit on the number of executors for all resource profiles.
   */
  private[deploy] def getExecutorLimit: Int = {
    targetNumExecutorsPerResourceProfileId.values.sum
  }

  def duration: Long = {
    if (endTime != -1) {
      endTime - startTime
    } else {
      System.currentTimeMillis() - startTime
    }
  }
}

相关信息

spark 源码目录

相关文章

spark ApplicationSource 源码

spark ApplicationState 源码

spark DriverInfo 源码

spark DriverState 源码

spark ExecutorDesc 源码

spark ExecutorResourceDescription 源码

spark FileSystemPersistenceEngine 源码

spark LeaderElectionAgent 源码

spark Master 源码

spark MasterArguments 源码

0  赞