| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134 | 
							- <!--
 
-   Licensed under the Apache License, Version 2.0 (the "License");
 
-   you may not use this file except in compliance with the License.
 
-   You may obtain a copy of the License at
 
-     http://www.apache.org/licenses/LICENSE-2.0
 
-   Unless required by applicable law or agreed to in writing, software
 
-   distributed under the License is distributed on an "AS IS" BASIS,
 
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 
-   See the License for the specific language governing permissions and
 
-   limitations under the License. See accompanying LICENSE file.
 
- -->
 
- <configuration>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.maximum-applications</name>
 
-     <value>10000</value>
 
-     <description>
 
-       Maximum number of applications that can be pending and running.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
 
-     <value>0.5</value>
 
-     <description>
 
-       Maximum percent of resources in the cluster which can be used to run 
 
-       application masters i.e. controls number of concurrent running
 
-       applications.5
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.resource-calculator</name>
 
-     <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
 
-     <description>
 
-       The ResourceCalculator implementation to be used to compare 
 
-       Resources in the scheduler.
 
-       The default i.e. DefaultResourceCalculator only uses Memory while
 
-       DominantResourceCalculator uses dominant-resource to compare 
 
-       multi-dimensional resources such as Memory, CPU etc.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.queues</name>
 
-     <value>default</value>
 
-     <description>
 
-       The queues at the this level (root is the root queue).
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.capacity</name>
 
-     <value>100</value>
 
-     <description>Default queue target capacity.</description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
 
-     <value>1</value>
 
-     <description>
 
-       Default queue user limit a percentage from 0.0 to 1.0.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
 
-     <value>100</value>
 
-     <description>
 
-       The maximum capacity of the default queue. 
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.state</name>
 
-     <value>RUNNING</value>
 
-     <description>
 
-       The state of the default queue. State can be one of RUNNING or STOPPED.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
 
-     <value>*</value>
 
-     <description>
 
-       The ACL of who can submit jobs to the default queue.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.root.default.acl_administer_queue</name>
 
-     <value>*</value>
 
-     <description>
 
-       The ACL of who can administer jobs on the default queue.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.node-locality-delay</name>
 
-     <value>40</value>
 
-     <description>
 
-       Number of missed scheduling opportunities after which the CapacityScheduler 
 
-       attempts to schedule rack-local containers. 
 
-       Typically this should be set to number of nodes in the cluster, By default is setting 
 
-       approximately number of nodes in one rack which is 40.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.queue-mappings</name>
 
-     <value></value>
 
-     <description>
 
-       A list of mappings that will be used to assign jobs to queues
 
-       The syntax for this list is [u|g]:[name]:[queue_name][,next mapping]*
 
-       Typically this list will be used to map users to queues,
 
-       for example, u:%user:%user maps all users to queues with the same name
 
-       as the user.
 
-     </description>
 
-   </property>
 
-   <property>
 
-     <name>yarn.scheduler.capacity.queue-mappings-override.enable</name>
 
-     <value>false</value>
 
-     <description>
 
-       If a queue mapping is present, will it override the value specified
 
-       by the user? This can be used by administrators to place jobs in queues
 
-       that are different than the one specified by the user.
 
-       The default is false.
 
-     </description>
 
-   </property>
 
- </configuration>
 
 
  |