Permalink
Cannot retrieve contributors at this time
Fetching contributors…
| --- | |
| ### This file implements defaults and some dependant parameter defaulting logic. | |
| ### Every parameter can be overridden using the hiera lookup hierarchy. The enclosd | |
| ### hiera.yaml provides for this by adding a site.yaml to the lookup where | |
| ### site-specific overrides can be placed. Therefore this file should never need | |
| ### changing by site admins. | |
| # FQDN of Namenode | |
| #bigtop::hadoop_head_node: "hadoopmaster.example.com" | |
| # FQDN of standby node (enables HA if set) | |
| #bigtop::hadoop_standby_head_node: "standbyNN.example.com" | |
| # FQDN of gateway node (if separate from NN) | |
| #bigtop::hadoop_gateway_node: "gateway.example.com" | |
| # A list of stack' components to be deployed can be specified via special | |
| # "$components" list. If $components isn't set then everything in the stack will | |
| # be installed as usual. Otherwise only a specified list will be set | |
| # Possible elements: | |
| # hadoop,yarn,hbase,alluxio,flink,flume,solrcloud,spark,oozie,hcat,sqoop,sqoop2,httpfs, | |
| # hue,mahout,giraph,crunch,pig,hive,zookeeper,ycsb,qfs | |
| # Example (to deploy only HDFS and YARN server and gateway parts) | |
| # This can be a comma-separated list or an array. | |
| #hadoop_cluster_node::cluster_components: | |
| # - hadoop | |
| # - yarn | |
| # A list of roles can be assigned to each node in a cluster via "bigtop::roles" | |
| # hiera conf. This is useful if you want to run specific daemons/processes of a component | |
| # (lets say hadoop) on specific nodes. To enable roles, you need to explicitly set | |
| # "bigtop::roles_enabled" conf to true. By default, it is not enabled. If not enabled, | |
| # bigtop will honour the hadoop_cluster_node::cluster_components parameter and assume | |
| # respective roles to each node based on hadoop_head_node, hadoop_gateway_node | |
| # and other nodes (worker nodes) classification. For full list of roles | |
| # supported by each component, refer to $roles_map datastructure in cluster.pp file | |
| # for example: | |
| #bigtop::roles: | |
| # - namenode | |
| # - resourcemanager | |
| # By specifying above roles, you instruct a node to run only namenode and resourcemanager. | |
| # When roles are disabled, all worker daemons for each component assume hadoop_head_node as | |
| # their master. With roles enabled, this assumption may no longer be valid and its upto the user | |
| # to specify the FQDN for each master daemons to all nodes so that worker daemons can communicate | |
| # with master. FQDN and port for master daemons of each component can be specified using below | |
| # configurations. | |
| # Storage directories (will be created if doesn't exist) | |
| #hadoop::hadoop_storage_dirs: | |
| # - /data/1 | |
| # - /data/2 | |
| # - /data/3 | |
| # - /data/4 | |
| #bigtop::bigtop_repo_uri: "http://mirror.example.com/path/to/mirror/" | |
| # Use a pre-installed java environment. The default value of 'false' will cause | |
| # the default jdk by bigtop to be installed. | |
| # this to 'true' requires a compatible java environment be avaialble prior to Bigtop | |
| # installation. | |
| #bigtop::jdk_preinstalled: false | |
| # Test-only variable controls if user hdfs' sshkeys should be installed to allow | |
| # for passwordless login across the cluster. Required by some integration tests | |
| #hadoop::common_hdfs::testonly_hdfs_sshkeys: "no" | |
| # Default | |
| #hadoop::common_hdfs::ha: "disabled" | |
| # Kerberos | |
| #hadoop::hadoop_security_authentication: "kerberos" | |
| #kerberos::site::domain: "do.main" | |
| #kerberos::site::realm: "DO.MAIN" | |
| #kerberos::site::kdc_server: "localhost" | |
| #kerberos::site::kdc_port: "88" | |
| #kerberos::site::admin_port: "749" | |
| #kerberos::site::keytab_export_dir: "/var/lib/bigtop_keytabs" | |
| # applies to hdfs, yarn, mapred and httpfs | |
| hadoop::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| hadoop::common_hdfs::hadoop_namenode_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| # actually default but needed for hadoop_namenode_uri here | |
| hadoop::common_hdfs::hadoop_namenode_port: "8020" | |
| # If there is a datanode/network failure in the write pipeline, DFSClient will try | |
| # to remove the failed datanode from the pipeline and then continue writing with the | |
| # remaining datanodes. As a result, the number of datanodes in the pipeline is decreased. | |
| # The feature is to add new datanodes to the pipeline. This is a site-wide property to | |
| # enable/disable the feature. When the cluster size is extremely small, e.g. 3 nodes or | |
| # less, cluster administrators may want to set the policy to NEVER in the default | |
| # configuration file or disable this feature. Otherwise, users may experience an unusually | |
| # high rate of pipeline failures since it is impossible to find new datanodes for replacement. | |
| # See also dfs.client.block.write.replace-datanode-on-failure.policy | |
| # hadoop::common_hdfs::hdfs_replace_datanode_on_failure: "false" | |
| # Set as shown below in site.yaml to also enable Kerberos authentication | |
| # on the web GUIs of journalnode, namenode, datanode, resourcemanager and | |
| # nodemanager when you enable Kerberos for Hadoop API communication. This | |
| # intentionally is not the default right now. | |
| #hadoop::common_hdfs::hadoop_http_authentication_type: "%{hiera('hadoop::hadoop_security_authentication')}" | |
| # | |
| # A secret is necessary for the cross-service-cross-node session cookie. | |
| # Provide this by setting the following to something long and secret: | |
| #hadoop::common_hdfs::hadoop_http_authentication_signature_secret: | |
| # ... or have trocla generate secrets for Hadoop components automatically. For | |
| # this to work, the trocla puppet module must be installed. | |
| #hadoop::generate_secrets: true | |
| hadoop::common_yarn::hadoop_ps_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| hadoop::common_yarn::hadoop_rm_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| # actually default but needed for hue::server::rm_port here | |
| hadoop::common_yarn::hadoop_rm_port: "8032" | |
| hadoop::common_mapred_app::jobtracker_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| hadoop::common_mapred_app::mapreduce_jobhistory_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| # actually default but needed for hadoop::common_yarn::yarn_log_server_url here | |
| bigtop::hadoop_history_server_port: "19888" | |
| bigtop::hadoop_history_server_url: "http://%{hiera('hadoop::common_mapred_app::mapreduce_jobhistory_host')}:%{hiera('bigtop::hadoop_history_server_port')}" | |
| hadoop::common_yarn::yarn_log_server_url: "%{hiera('bigtop::hadoop_history_server_url')}/jobhistory/logs" | |
| # actually default but needed for hue::server::webhdfs_url here | |
| hadoop::httpfs::hadoop_httpfs_port: "14000" | |
| bigtop::hadoop_zookeeper_port: "2181" | |
| hadoop::zk: "%{hiera('bigtop::hadoop_head_node')}:%{hiera('bigtop::hadoop_zookeeper_port')}" | |
| bigtop::hadoop_namenode_uri: "hdfs://%{hiera('hadoop::common_hdfs::hadoop_namenode_host')}:%{hiera('hadoop::common_hdfs::hadoop_namenode_port')}" | |
| hadoop_hbase::base_relative_rootdir: "/hbase" | |
| hadoop_hbase::common_config::rootdir: "%{hiera('bigtop::hadoop_namenode_uri')}%{hiera('hadoop_hbase::base_relative_rootdir')}" | |
| hadoop_hbase::common_config::zookeeper_quorum: "%{hiera('bigtop::hadoop_head_node')}" | |
| hadoop_hbase::common_config::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| hadoop_hbase::client::thrift: true | |
| hadoop_hbase::deploy::auxiliary: true | |
| solr::server::root_url: "%{hiera('bigtop::hadoop_namenode_uri')}" | |
| solr::server::zk: "%{hiera('hadoop::zk')}" | |
| solr::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| # Default but needed here to make sure, hue uses the same port | |
| solr::server::port: "8983" | |
| hadoop_oozie::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| hcatalog::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| hcatalog::webhcat::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| # spark | |
| spark::common::master_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| # to enable spark HA, ensure zookeeper is available and uncomment the line below | |
| #spark::common::zookeeper_connection_string: "%{hiera('hadoop::zk')}" | |
| alluxio::common::master_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| # qfs | |
| qfs::common::metaserver_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| qfs::common::metaserver_port: "30000" | |
| qfs::common::chunkserver_port: "30000" | |
| qfs::common::metaserver_client_port: "20000" | |
| qfs::common::chunkserver_client_port: "22000" | |
| hadoop_zookeeper::server::myid: "0" | |
| hadoop_zookeeper::server::ensemble: | |
| - ["0", "%{hiera('bigtop::hadoop_head_node')}:2888:3888"] | |
| hadoop_zookeeper::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| # those are only here because they were present as extlookup keys previously | |
| bigtop::hadoop_rm_http_port: "8088" | |
| bigtop::hadoop_rm_proxy_port: "8088" | |
| bigtop::sqoop2_server_port: "12000" | |
| bigtop::hbase_thrift_port: "9090" | |
| bigtop::hadoop_oozie_port: "11000" | |
| hue::server::rm_host: "%{hiera('hadoop::common_yarn::hadoop_rm_host')}" | |
| hue::server::rm_port: "%{hiera('hadoop::common_yarn::hadoop_rm_port')}" | |
| hue::server::rm_url: "http://%{hiera('bigtop::hadoop_head_node')}:%{hiera('bigtop::hadoop_rm_http_port')}" | |
| hue::server::rm_proxy_url: "http://%{hiera('hadoop::common_yarn::hadoop_ps_host')}:%{hiera('hadoop::common_yarn::hadoop_ps_port')}" | |
| hue::server::history_server_url: "%{hiera('bigtop::hadoop_history_server_url')}" | |
| # those use fqdn instead of hadoop_head_node because it's only ever activated | |
| # on the gatewaynode | |
| hue::server::webhdfs_url: "http://%{fqdn}:%{hiera('hadoop::httpfs::hadoop_httpfs_port')}/webhdfs/v1" | |
| hue::server::sqoop2_url: "http://%{fqdn}:%{hiera('bigtop::sqoop2_server_port')}/sqoop" | |
| hue::server::solr_url: "http://%{fqdn}:%{hiera('solr::server::port')}/solr/" | |
| hue::server::hbase_thrift_url: "%{fqdn}:%{hiera('bigtop::hbase_thrift_port')}" | |
| hue::server::oozie_url: "http://%{hiera('bigtop::hadoop_head_node')}:%{hiera('bigtop::hadoop_oozie_port')}/oozie" | |
| hue::server::default_fs: "%{hiera('bigtop::hadoop_namenode_uri')}" | |
| hue::server::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| giraph::client::zookeeper_quorum: "%{hiera('bigtop::hadoop_head_node')}" | |
| hadoop_hive::common_config::hbase_zookeeper_quorum: "%{hiera('hadoop_hbase::common_config::zookeeper_quorum')}" | |
| hadoop_hive::common_config::kerberos_realm: "%{hiera('kerberos::site::realm')}" | |
| hadoop_hive::common_config::metastore_uris: "thrift://%{hiera('bigtop::hadoop_head_node')}:9083" | |
| # tez | |
| hadoop::common::tez_conf_dir: "/etc/tez/conf" | |
| hadoop::common::tez_jars: "/usr/lib/tez" | |
| # to enable tez in hadoop, uncomment the lines below | |
| # hadoop::common::use_tez: true | |
| # hadoop::common_mapred_app::mapreduce_framework_name: "yarn-tez" | |
| # to enable tez in hive, uncomment the lines below | |
| # hadoop_hive::common_config::hive_execution_engine: "tez" | |
| #kafka | |
| kafka::server::port: "9092" | |
| kafka::server::zookeeper_connection_string: "%{hiera('bigtop::hadoop_head_node')}:2181" | |
| zeppelin::server::spark_master_url: "yarn-client" | |
| zeppelin::server::hiveserver2_url: "jdbc:hive2://%{hiera('hadoop-hive::common::hiveserver2_host')}:%{hiera('hadoop-hive::common::hiveserver2_port')}" | |
| zeppelin::server::hiveserver2_user: "%{hiera('bigtop::hiveserver2_user')}" | |
| zeppelin::server::hiveserver2_password: "%{hiera('bigtop::hiveserver2_password')}" | |
| # Flink | |
| flink::common::jobmanager_host: "%{hiera('bigtop::hadoop_head_node')}" | |
| flink::common::jobmanager_port: "6123" | |
| flink::common::ui_port: "8081" | |
| flink::common::storage_dirs: "%{hiera('hadoop::hadoop_storage_dirs')}" | |
| gpdb::common::nodes: ["master_node_fqdn", "segment_node_fqdn"] | |
| gpdb::common::gp_home: "/usr/lib/gpdb" | |
| gpdb::common::db_base_dir: "/data_gp" | |
| gpdb::common::master_db_port: "5432" | |
| gpdb::common::segment_db_port_prefix: "4000" | |
| ambari::agent::server_host: "%{hiera('bigtop::hadoop_head_node')}" |