diff --git a/data/templates/telegraf/telegraf.j2 b/data/templates/telegraf/telegraf.j2
index f382dbf2e..535e3a347 100644
--- a/data/templates/telegraf/telegraf.j2
+++ b/data/templates/telegraf/telegraf.j2
@@ -1,139 +1,141 @@
 # Generated by /usr/libexec/vyos/conf_mode/service_monitoring_telegraf.py
 
 [agent]
   interval = "15s"
   round_interval = true
   metric_batch_size = 1000
   metric_buffer_limit = 10000
   collection_jitter = "5s"
   flush_interval = "15s"
   flush_jitter = "0s"
   precision = ""
   debug = false
   quiet = false
   logfile = ""
   hostname = "{{ hostname }}"
   omit_hostname = false
 {% if azure_data_explorer is vyos_defined %}
 ### Azure Data Explorer ###
 [[outputs.azure_data_explorer]]
   ## The URI property of the Azure Data Explorer resource on Azure
   endpoint_url = "{{ azure_data_explorer.url }}"
 
   ## The Azure Data Explorer database that the metrics will be ingested into.
   ## The plugin will NOT generate this database automatically, it's expected that this database already exists before ingestion.
   database = "{{ azure_data_explorer.database }}"
   metrics_grouping_type = "{{ azure_data_explorer.group_metrics }}"
 
   ## Name of the single table to store all the metrics (Only needed if metrics_grouping_type is "SingleTable").
 {%     if azure_data_explorer.table is vyos_defined and azure_data_explorer.group_metrics == 'SingleTable' %}
   table_name = "{{ azure_data_explorer.table }}"
 {%     endif %}
 ### End Azure Data Explorer ###
 {% endif %}
 {% if influxdb is vyos_defined %}
 ### InfluxDB2 ###
 [[outputs.influxdb_v2]]
   urls = ["{{ influxdb.url }}:{{ influxdb.port }}"]
   insecure_skip_verify = true
   token = "$INFLUX_TOKEN"
   organization = "{{ influxdb.authentication.organization }}"
   bucket = "{{ influxdb.bucket }}"
 ### End InfluxDB2 ###
 {% endif %}
 {% if loki is vyos_defined %}
 ### Loki ###
 [[outputs.loki]]
   ## The domain of Loki
   domain = "{{ loki.url }}:{{ loki.port }}"
 {%     if loki.authentication.username is vyos_defined and loki.authentication.password is vyos_defined  %}
   ## Basic Authentication
   username = "{{ loki.authentication.username }}"
   password = "{{ loki.authentication.password }}"
 {%     endif %}
 {%     if loki.metric_name_label is vyos_defined %}
 metric_name_label = "{{ loki.metric_name_label }}"
 {%     endif %}
 ### End Loki ###
 {% endif %}
 {% if prometheus_client is vyos_defined %}
 ### Prometheus ###
 [[outputs.prometheus_client]]
   ## Address to listen on
   listen = "{{ prometheus_client.listen_address | bracketize_ipv6 if prometheus_client.listen_address is vyos_defined else '' }}:{{ prometheus_client.port }}"
   metric_version = {{ prometheus_client.metric_version }}
 {%     if prometheus_client.authentication.username is vyos_defined and prometheus_client.authentication.password is vyos_defined  %}
   ## Use HTTP Basic Authentication
   basic_username = "{{ prometheus_client.authentication.username }}"
   basic_password = "{{ prometheus_client.authentication.password }}"
 {%     endif %}
 {%     if prometheus_client.allow_from is vyos_defined %}
   ip_range = {{ prometheus_client.allow_from }}
 {%     endif %}
 ### End Prometheus ###
 {% endif %}
 {% if splunk is vyos_defined %}
 ### Splunk ###
 [[outputs.http]]
   ## URL is the address to send metrics to
   url = "{{ splunk.url }}"
   ## Timeout for HTTP message
   # timeout = "5s"
   ## Use TLS but skip chain & host verification
 {%     if splunk.authentication.insecure is vyos_defined %}
   insecure_skip_verify = true
 {%     endif %}
   ## Data format to output
   data_format = "splunkmetric"
   ## Provides time, index, source overrides for the HEC
   splunkmetric_hec_routing = true
   ## Additional HTTP headers
    [outputs.http.headers]
    # Should be set manually to "application/json" for json data_format
      Content-Type = "application/json"
      Authorization = "Splunk {{ splunk.authentication.token }}"
      X-Splunk-Request-Channel = "{{ splunk.authentication.token }}"
 ### End Splunk ###
 {% endif %}
 [[inputs.cpu]]
     percpu = true
     totalcpu = true
     collect_cpu_time = false
     report_active = false
 [[inputs.disk]]
     ignore_fs = ["devtmpfs", "devfs"]
 [[inputs.diskio]]
 [[inputs.mem]]
 [[inputs.net]]
     ignore_protocol_stats = true
 [[inputs.nstat]]
 [[inputs.system]]
 [[inputs.netstat]]
 [[inputs.processes]]
 [[inputs.kernel]]
 [[inputs.interrupts]]
 [[inputs.linux_sysctl_fs]]
 [[inputs.systemd_units]]
 [[inputs.conntrack]]
   files = ["ip_conntrack_count","ip_conntrack_max","nf_conntrack_count","nf_conntrack_max"]
   dirs = ["/proc/sys/net/ipv4/netfilter","/proc/sys/net/netfilter"]
 [[inputs.ethtool]]
   interface_include = {{ interfaces_ethernet }}
 [[inputs.chrony]]
   dns_lookup = true
 [[inputs.internal]]
 [[inputs.nstat]]
 [[inputs.syslog]]
   server = "unixgram:///run/telegraf/telegraf_syslog.sock"
   best_effort = true
   syslog_standard = "RFC3164"
 {% if influxdb is vyos_defined %}
 [[inputs.exec]]
   commands = [
+{%     if nft_chains is vyos_defined %}
     "{{ custom_scripts_dir }}/show_firewall_input_filter.py",
+{%     endif %}
     "{{ custom_scripts_dir }}/show_interfaces_input_filter.py",
     "{{ custom_scripts_dir }}/vyos_services_input_filter.py"
   ]
   timeout = "10s"
   data_format = "influx"
 {% endif %}
diff --git a/src/conf_mode/service_monitoring_telegraf.py b/src/conf_mode/service_monitoring_telegraf.py
index 9455b6109..db870aae5 100755
--- a/src/conf_mode/service_monitoring_telegraf.py
+++ b/src/conf_mode/service_monitoring_telegraf.py
@@ -1,237 +1,238 @@
 #!/usr/bin/env python3
 #
 # Copyright (C) 2021-2024 VyOS maintainers and contributors
 #
 # This program is free software; you can redistribute it and/or modify
 # it under the terms of the GNU General Public License version 2 or later as
 # published by the Free Software Foundation.
 #
 # This program is distributed in the hope that it will be useful,
 # but WITHOUT ANY WARRANTY; without even the implied warranty of
 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 # GNU General Public License for more details.
 #
 # You should have received a copy of the GNU General Public License
 # along with this program.  If not, see <http://www.gnu.org/licenses/>.
 
 import os
 import socket
 import json
 
 from sys import exit
 from shutil import rmtree
 
 from vyos.config import Config
 from vyos.configdict import is_node_changed
 from vyos.configverify import verify_vrf
 from vyos.ifconfig import Section
 from vyos.template import render
 from vyos.utils.process import call
 from vyos.utils.permission import chown
 from vyos.utils.process import cmd
 from vyos import ConfigError
 from vyos import airbag
 airbag.enable()
 
 cache_dir = f'/etc/telegraf/.cache'
 config_telegraf = f'/run/telegraf/telegraf.conf'
 custom_scripts_dir = '/etc/telegraf/custom_scripts'
 syslog_telegraf = '/etc/rsyslog.d/50-telegraf.conf'
 systemd_override = '/run/systemd/system/telegraf.service.d/10-override.conf'
 
 def get_nft_filter_chains():
     """ Get nft chains for table filter """
     try:
         nft = cmd('nft --json list table ip vyos_filter')
     except Exception:
         print('nft table ip vyos_filter not found')
         return []
     nft = json.loads(nft)
     chain_list = []
 
     for output in nft['nftables']:
         if 'chain' in output:
             chain = output['chain']['name']
             chain_list.append(chain)
 
     return chain_list
 
 def get_hostname() -> str:
     try:
         hostname = socket.getfqdn()
     except socket.gaierror:
         hostname = socket.gethostname()
     return hostname
 
 def get_config(config=None):
     if config:
         conf = config
     else:
         conf = Config()
     base = ['service', 'monitoring', 'telegraf']
     if not conf.exists(base):
         return None
 
     monitoring = conf.get_config_dict(base, key_mangling=('-', '_'),
                                       get_first_key=True,
                                       no_tag_node_value_mangle=True)
 
     tmp = is_node_changed(conf, base + ['vrf'])
     if tmp: monitoring.update({'restart_required': {}})
 
     # We have gathered the dict representation of the CLI, but there are default
     # options which we need to update into the dictionary retrived.
     monitoring = conf.merge_defaults(monitoring, recursive=True)
 
     monitoring['custom_scripts_dir'] = custom_scripts_dir
     monitoring['hostname'] = get_hostname()
     monitoring['interfaces_ethernet'] = Section.interfaces('ethernet', vlan=False)
-    monitoring['nft_chains'] = get_nft_filter_chains()
+    if conf.exists('firewall'):
+        monitoring['nft_chains'] = get_nft_filter_chains()
 
     # Redefine azure group-metrics 'single-table' and 'table-per-metric'
     if 'azure_data_explorer' in monitoring:
         if 'single-table' in monitoring['azure_data_explorer']['group_metrics']:
             monitoring['azure_data_explorer']['group_metrics'] = 'SingleTable'
         else:
             monitoring['azure_data_explorer']['group_metrics'] = 'TablePerMetric'
         # Set azure env
         if 'authentication' in monitoring['azure_data_explorer']:
             auth_config = monitoring['azure_data_explorer']['authentication']
             if {'client_id', 'client_secret', 'tenant_id'} <= set(auth_config):
                 os.environ['AZURE_CLIENT_ID'] = auth_config['client_id']
                 os.environ['AZURE_CLIENT_SECRET'] = auth_config['client_secret']
                 os.environ['AZURE_TENANT_ID'] = auth_config['tenant_id']
 
     # Ignore default XML values if config doesn't exists
     # Delete key from dict
     if not conf.exists(base + ['influxdb']):
         del monitoring['influxdb']
 
     if not conf.exists(base + ['prometheus-client']):
         del monitoring['prometheus_client']
 
     if not conf.exists(base + ['azure-data-explorer']):
         del monitoring['azure_data_explorer']
 
     if not conf.exists(base + ['loki']):
         del monitoring['loki']
 
     return monitoring
 
 def verify(monitoring):
     # bail out early - looks like removal from running config
     if not monitoring:
         return None
 
     verify_vrf(monitoring)
 
     # Verify influxdb
     if 'influxdb' in monitoring:
         if 'authentication' not in monitoring['influxdb'] or \
            'organization' not in monitoring['influxdb']['authentication'] or \
            'token' not in monitoring['influxdb']['authentication']:
             raise ConfigError(f'influxdb authentication "organization and token" are mandatory!')
 
         if 'url' not in monitoring['influxdb']:
             raise ConfigError(f'Monitoring influxdb "url" is mandatory!')
 
     # Verify azure-data-explorer
     if 'azure_data_explorer' in monitoring:
         if 'authentication' not in monitoring['azure_data_explorer'] or \
            'client_id' not in monitoring['azure_data_explorer']['authentication'] or \
            'client_secret' not in monitoring['azure_data_explorer']['authentication'] or \
            'tenant_id' not in monitoring['azure_data_explorer']['authentication']:
             raise ConfigError(f'Authentication "client-id, client-secret and tenant-id" are mandatory!')
 
         if 'database' not in monitoring['azure_data_explorer']:
             raise ConfigError(f'Monitoring "database" is mandatory!')
 
         if 'url' not in monitoring['azure_data_explorer']:
             raise ConfigError(f'Monitoring "url" is mandatory!')
 
         if monitoring['azure_data_explorer']['group_metrics'] == 'SingleTable' and \
             'table' not in monitoring['azure_data_explorer']:
              raise ConfigError(f'Monitoring "table" name for single-table mode is mandatory!')
 
     # Verify Splunk
     if 'splunk' in monitoring:
         if 'authentication' not in monitoring['splunk'] or \
            'token' not in monitoring['splunk']['authentication']:
             raise ConfigError(f'Authentication "organization and token" are mandatory!')
 
         if 'url' not in monitoring['splunk']:
             raise ConfigError(f'Monitoring splunk "url" is mandatory!')
 
     # Verify Loki
     if 'loki' in monitoring:
         if 'url' not in monitoring['loki']:
             raise ConfigError(f'Monitoring loki "url" is mandatory!')
         if 'authentication' in monitoring['loki']:
             if (
                 'username' not in monitoring['loki']['authentication']
                 or 'password' not in monitoring['loki']['authentication']
             ):
                 raise ConfigError(
                     f'Authentication "username" and "password" are mandatory!'
                 )
 
     return None
 
 def generate(monitoring):
     if not monitoring:
         # Delete config and systemd files
         config_files = [config_telegraf, systemd_override, syslog_telegraf]
         for file in config_files:
             if os.path.isfile(file):
                 os.unlink(file)
 
         # Delete old directories
         if os.path.isdir(cache_dir):
             rmtree(cache_dir, ignore_errors=True)
 
         return None
 
     # Create telegraf cache dir
     if not os.path.exists(cache_dir):
         os.makedirs(cache_dir)
 
     chown(cache_dir, 'telegraf', 'telegraf')
 
     # Create custome scripts dir
     if not os.path.exists(custom_scripts_dir):
         os.mkdir(custom_scripts_dir)
 
     # Render telegraf configuration and systemd override
     render(config_telegraf, 'telegraf/telegraf.j2', monitoring, user='telegraf', group='telegraf')
     render(systemd_override, 'telegraf/override.conf.j2', monitoring)
     render(syslog_telegraf, 'telegraf/syslog_telegraf.j2', monitoring)
 
     return None
 
 def apply(monitoring):
     # Reload systemd manager configuration
     systemd_service = 'telegraf.service'
     call('systemctl daemon-reload')
     if not monitoring:
         call(f'systemctl stop {systemd_service}')
         return
 
     # we need to restart the service if e.g. the VRF name changed
     systemd_action = 'reload-or-restart'
     if 'restart_required' in monitoring:
         systemd_action = 'restart'
 
     call(f'systemctl {systemd_action} {systemd_service}')
 
     # Telegraf include custom rsyslog config changes
     call('systemctl reload-or-restart rsyslog')
 
 if __name__ == '__main__':
     try:
         c = get_config()
         verify(c)
         generate(c)
         apply(c)
     except ConfigError as e:
         print(e)
         exit(1)