Get a yarn configuration from commandline

前端 未结 1 1913
长发绾君心
长发绾君心 2020-12-31 06:41

In EMR, is there a way to get a specific value of the configuration given the configuration key using the yarn command?

For example I would like to do s

相关标签:
1条回答
  • 2020-12-31 07:40

    It's a bit non-intuitive, but it turns out the hdfs getconf command is capable of checking configuration properties for YARN and MapReduce, not only HDFS.

    > hdfs getconf -confKey fs.defaultFS
    hdfs://localhost:19000
    
    > hdfs getconf -confKey dfs.namenode.name.dir
    file:///Users/chris/hadoop-deploy-trunk/data/dfs/name
    
    > hdfs getconf -confKey yarn.resourcemanager.address
    0.0.0.0:8032
    
    > hdfs getconf -confKey mapreduce.framework.name
    yarn
    

    A benefit of using this is that you'll see the actual, final results of any configuration properties as they are actually used by Hadoop. This would account for some of the more advanced configuration patterns, such as use of XInclude in the XML files or property substitutions, like this:

      <property>
        <description>The address of the applications manager interface in the RM.</description>
        <name>yarn.resourcemanager.address</name>
        <value>${yarn.resourcemanager.hostname}:8032</value>
      </property>
    

    Any scripting approach that tries to parse the XML files directly is unlikely to accurately match the implementation as its done inside Hadoop, so it's better to ask Hadoop itself.

    You might be wondering why an hdfs command can get configuration properties for YARN and MapReduce. Great question! It's somewhat of a coincidence of the implementation needing to inject an instance of MapReduce's JobConf into some objects created via reflection. The relevant code is visible here:

    https://github.com/apache/hadoop/blob/release-2.7.1/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ReflectionUtils.java#L82-L114

    This code is executed as part of running the hdfs getconf command. By triggering a reference to JobConf, it forces class loading and static initialization of the relevant MapReduce and YARN classes that add yarn-default.xml, yarn-site.xml, mapred-default.xml and mapred-site.xml to the set of configuration files in effect.

    Since it's a coincidence of the implementation, it's possible that some of this behavior will change in future versions, but it would be a backwards-incompatible change, so we definitely wouldn't change that behavior inside the current Hadoop 2.x line. The Apache Hadoop Compatibility policy commits to backwards-compatibility within a major version line, so you can trust that this will continue working at least within the 2.x version line.

    0 讨论(0)
提交回复
热议问题