• 地址信息:xxx主机上集成到ambari的prometheus
  • /opt/prometheus/prometheus-2.34.0.linux-amd64
  • # my global config
  • global:
  • scrape_interval: 25s # Set the scrape interval to every 15 seconds. Default is every 1 minute.
  • evaluation_interval: 25s # Evaluate rules every 15 seconds. The default is every 1 minute.
  • # scrape_timeout is set to the global default (10s).
  • # Alertmanager configuration
  • alerting:
  • alertmanagers:
  • – static_configs:
  • – targets:
  • # – alertmanager:9093
  • # Load rules once and periodically evaluate them according to the global ‘evaluation_interval’.
  • rule_files:
  • # – “first_rules.yml”
  • # – “second_rules.yml”
  • # A scrape configuration containing exactly one endpoint to scrape:
  • # Here it’s Prometheus itself.
  • scrape_configs:
  • # The job name is added as a label `job=` to any timeseries scraped from this config.
  • – job_name: ‘prometheus’
  • # metrics_path defaults to ‘/metrics’
  • # scheme defaults to ‘http’.
  • static_configs:
  • – targets: [‘hadoop001:9090’]
  • – job_name: ‘node_export_hadoop01’
  • static_configs:
  • – targets: [‘hadoop01:9100’]
  • – job_name: ‘node_export_hadoop02’
  • static_configs:
  • – targets: [‘hadoop02:9100’]
  • – job_name: ‘node_export_hadoop03’
  • static_configs:
  • – targets: [‘hadoop03:9100’]

作者 admin

张宴银,大数据开发工程师

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注