1. Official Apache Spark Documentation: The documentation explicitly states the purpose of spark.executor.cores.
Source: Apache Spark 3.4.1 Documentation
Configuration Page
Application Properties section.
Reference: Under the property spark.executor.cores
the description reads: "The number of cores to use on each executor. In standalone and YARN mode
this controls the number of concurrent tasks an executor can run."
2. Official Databricks Documentation: Databricks documentation reinforces this concept in its cluster configuration guides.
Source: Databricks Documentation
Cluster configuration best practices.
Reference: In discussions about cluster sizing and tuning
it is explained that the number of executor cores determines the number of task slots available per executor
directly impacting parallelism. For example
the "Worker type" section often links the number of vCPUs to the default spark.executor.cores setting.
3. University Courseware: Academic materials on distributed computing with Spark confirm this fundamental architectural principle.
Source: UC Berkeley
CS 186/286
Introduction to Database Systems (Fall 2018)
Lecture 20: Spark.
Reference: Slide 25 ("Anatomy of a Spark Job") illustrates that an Executor contains multiple "Task Slots" (cores)
and each slot runs one task. The number of these slots is configured by spark.executor.cores.