赞
踩
工作节点:HF OR Indexer
NO Indexer Cluster
流程:
props -> transforms -> outputs
props.conf transforms.conf outputs.conf 确定哪些数据进行数据转发 配置数据转发策略 配置目标ip端口消息
outputs.conf
# 默认数据转发组
[syslog]
defaultGroup = syslogoutputsGroup
# 配置syslog数据转发目标地址
[syslog:syslogoutputsGroup]
type = udp
# 协议
server = 10.203.19.22:514
# 目标ip和端口
props.conf
# 需要转发的sourcetype名称
[sourcetype::cosmo_syslog]
TRANSFORMS-syslog = splunk_syslog_outpust
# 需要转发的sourcetype名称
[sourcetype::cosmo_cisco:asa]
TRANSFORMS-cisco-asa = splunk_syslog_outpust
# 需要转发的sourcetype名称
[sourcetype::cosmo_cisco_asa_anquan]
TRANSFORMS-cisco_asa_anquan] = splunk_syslog_outpust
transforms.conf
[splunk_syslog_outpust]
REGEX = .
DEST_KEY = _SYSLOG_ROUTING
FORMAT = syslogoutputsGroup
工作节点:HF OR Indexer
NO Indexer Cluster
设置transforms.conf创建过滤数据规则
问题:将普罗米修斯数据接入到splunk系统中发现数据量非常庞大(有1560个指标项),真正使用到的指标项不到100个,需要过滤掉不需要的指标
props.conf
[prometheus_filter]
# prometheus_filter sourcetype 解析规则
TIME_FORMAT = %s%3N
TIME_PREFIX = }\s[\d\-\.]+\s
NO_BINARY_CHECK = true
SHOULD_LINEMERGE = false
pulldown_type = 1
category = Metrics
EXTRACT-metric_name,num1,num2 = ^(?P<metric_name>[^\{]+)[^\}\n]*\}\s+(?P<num1>[^ ]+)\s+(?P<num2>.+)
# 指定transforms.conf 配置内容
TRANSFORMS-set = setnull,conform
transforms.conf
[conform]
REGEX = container_cpu_load_average_10s|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_usage_bytes|container_memory_cache|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_spec_memory_limit_bytes|jvm_gc_collection_seconds_count|jvm_gc_collection_seconds_sum|jvm_memory_bytes_used|jvm_memory_pool_bytes_max|jvm_memory_pool_bytes_used|jvm_threads_current|jvm_threads_deadlocked|jvm_threads_peak|kube_pod_container_status_ready|kube_pod_container_status_running|kube_pod_container_status_terminated|kube_pod_container_status_waiting|machine_cpu_cores|machine_memory_bytes|mysql_global_status_aborted_clients|mysql_global_status_commands_total|mysql_global_status_innodb_data_fsyncs|mysql_global_status_innodb_data_reads|mysql_global_status_innodb_data_writes|mysql_global_status_select_full_join|mysql_global_status_select_scan|mysql_global_status_slow_queries|mysql_global_status_sort_scan|mysql_global_status_threads_connected|mysql_global_status_threads_created|nginx_connections_accepted|nginx_connections_active|nginx_connections_handled|nginx_http_requests_total|node_network_info|rabbitmq_channelsTotal|rabbitmq_connectionsTotal|rabbitmq_consumersTotal|rabbitmq_exchangesTotal|rabbitmq_fd_total|rabbitmq_fd_used|rabbitmq_queue_messages_ready_total|rabbitmq_queue_messages_total|rabbitmq_queue_messages_unacknowledged_total|rabbitmq_queuesTotal|rabbitmq_sockets_total|rabbitmq_sockets_used|rabbitmq_up|redis_blocked_clients|redis_commands_processed_total|redis_connected_clients|redis_evicted_keys_total|redis_expired_keys_total|redis_keyspace_hits_total|redis_keyspace_misses_total|redis_memory_used_bytes|zk_max_latency|zk_num_alive_connections|zk_open_file_descriptor_count|zk_outstanding_requests|zk_up|zk_watch_count|zk_znode_count
# 需要采集的数据,通过正则表达式匹配。
# 内容为_raw内容 math(_raw,REGEX)
DEST_KEY = queue
FORMAT = indexQueue
# 转发到索引队列
[setnull]
REGEX = .
DEST_KEY = queue
FORMAT = nullQueue # 传递到nullQueue
实现原理:forwarder采集数据发送到HeavyForwarder上进行解析,由此在HeavyForwarder创建数据过滤规则。
工作节点:HF OR Indexer
NO Indexer Cluster
编辑props.conf
使用
SSH
工具登录到Heavy Forwarder服务器编辑
/opt/splunk/etc/apps/Splunk_TA_windows/local/props.conf
(如文件不存在则新建)。在文件末尾添加以下内容:
[host::example_hostname] # example-hostname 为排除数据的agent主机名 TRANSFORMS-filterhostname = conform,setnull [host::example_ip] # example-IP 为排除数据的agentIP地址 TRANSFORMS-filterIP = conform,setnull
- 1
- 2
- 3
- 4
- 5
编辑transforms.conf
编辑
/opt/splunk/etc/apps/Splunk_TA_windows/local/transforms.conf
(如文件不存在则新建)。在文件末尾添加以下内容:
[conform] REGEX = . DEST_KEY = queue FORMAT = indexQueue [setnull] # 数据不保存,转发到空消息队列 REGEX = ((\<EventID\>4703\<\/EventID\>)|(EventCode=4703)) DEST_KEY = queue FORMAT = nullQueue
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
访问HF refresh web界面刷新配置文件
http://HeavyForwarderIP:8000/zh-CN/debug/refresh
# 需要使用admin账号登入
界面如下:
完成效果如下:
不同种类的网络安全数据全部通过udp 514端口进行数据接收,创建采集配置无法指定多个sourcetype,可以根据source字段进行sourcetype拆分,匹配正则表达式拆分sourcetype
工作节点:HF OR Indexer
NO Indexer Cluster
props.conf
[source::example_udp:514]
SHOULD_LINEMERGE = false
TRANSFORMS-changesourcetype = mcafee_ips:syslog,f5_bigip:syslog,asiainfo:tda
transforms.conf
[asiainfo:tda]
REGEX = CEF:\d\|(Asiainfo Security|AsiaInfo|Trend Micro)\|(Control Manager|Deep Discovery Analyzer|Deep Security Agent|Deep Security Manager|TDA|Deep Discovery Email Inspector)
FORMAT = sourcetype::asiainfo:tda
DEST_KEY = MetaData:Sourcetype
[f5_bigip:syslog]
REGEX = \tASM\:CEF\:\d|F5|ASM
FORMAT = sourcetype::f5_bigip:syslog
DEST_KEY = MetaData:Sourcetype
[mcafee_ips:syslog]
REGEX = SyslogAlertForwarder:\s\|\d*
FORMAT = sourcetype::mcafee:ips:syslog
DEST_KEY = MetaData:Sourcetype
配置文件所在节点即工作节点:HF
props.conf
[splunktcp]
TRANSFORMS-routing = route_all
outputs.conf
# 注意点:不要加defaultGroup配置,否则数据优先给defaultGroup组发送。 # [tcpout] # defaultGroup = idxc_dv # 普通数据转发 [tcpout:mss_idx] disabled = false server = 10.10.0.104:9997 # 索引器发现 [tcpout:idxc_dv] indexerDiscovery = indexer_cluster_discovery [indexer_discovery:indexer_cluster_discovery] # 密码:idx_dv pass4SymmKey = $7$Vys7w3j3T5J1N0u5fr9JYbHGQe5Zh7Hef65c/I4vUOj8THEmBJI= master_uri = https://10.10.0.121:8089
补充点:
master
节点上面配置索引器发现
配置文件:$SPLUNK_HOME/etc/system/local/server.conf
[indexer_discovery]
# 密码:idx_dv
pass4SymmKey = $7$8OOMpbWXJscp69HifzghTxen+OgyLkvOkf51+1ioUDf0cyCjfRw=
polling_rate = 10
indexerWeightByDiskCapacity = false
transforms.conf
[route_all]
REGEX=(.)
DEST_KEY=_TCP_ROUTING
FORMAT=idxc_dv,mss_idx
用于转发HF节点采集syslog落盘数据
工作节点:HF
outputs.conf
[tcpout:mss_idxc]
server=server1:9997,server2:9997
[tcpout:local_idxc]
server=server3:9997,server4:9997
# 注意点:不要加defaultGroup配置,否则不生效
# [tcpout]
# defaultGroup = primary_indexers
inputs.conf
[monitor://.../file1.log]
_TCP_ROUTING = local_idxc
[monitor://.../file2.log]
_TCP_ROUTING = mss_idxc
[monitor://.../file3.log]
_TCP_ROUTING = local_idxc,mss_idxc
# 也可以配置多个outputs转发
查看/opt/splunkforwarder/etc/system/local/web.conf文件是否存在
[root@linux_mysql local]# ll /opt/splunkforwarder/etc/system/local/
web.conf
-rw-r--r--. 1 root root 52101 Jul 4 00:07 /opt/splunkforwarder/etc/system/local/web.conf
[root@linux_mysql local]#
如果web.conf文件不存在将…/default/下面的web.conf 复制到local目录下
cd /opt/splunkforwarder/etc/system/local/
cp ../default/web.conf .
找到mgmtHostPort配置项并修改
# Version 7.2.5 # DO NOT EDIT THIS FILE! # Changes to default files will be lost on update and are difficult to # manage and support. # # Please make any changes to system defaults by overriding them in # apps or $SPLUNK_HOME/etc/system/local # (See "Configuration file precedence" in the web documentation). # # To override a specific setting, copy the name of the stanza and # setting to the file where you wish to override it. # # This file contains possible attributes and values you can use to configure Splunk's web interface. # [default] [settings] # enable/disable the appserver startwebserver = 1 # port number tag is missing or 0 the server will NOT start an http listener # this is the port used for both SSL and non-SSL (we only have 1 port now). httpport = 8000 # this determines whether to start SplunkWeb in http or https. enableSplunkWebSSL = false # location of splunkd; don't include http[s]:// in this anymore. #mgmtHostPort = 127.0.0.1:8089--此项配置 mgmtHostPort = 127.0.0.1:8099 # list of ports to start python application servers on (although usually # one port is enough) Set to 0 to instead run the application server # directly as the web front end on 'httpport', separate from splunkd. appServerPorts = 8065 # default timeout, in seconds, when communicating with splunkd splunkdConnectionTimeout = 30 # enable/disable custom netloc when using http client enableSplunkWebClientNetloc = False # SSL certificate files. privKeyPath = $SPLUNK_HOME/etc/auth/splunkweb/privkey.pem serverCert = $SPLUNK_HOME/etc/auth/splunkweb/cert.pem
/opt/splunkforwarder/bin/splunk start
ps -ef|grep splunk
[root@linux_mysql local]# ps -ef|grep splunk
root 7698 1 0 00:10 ? 00:00:01 splunkd --under-systemd --systemd-delegate=yes -p 8099 _internal_launch_under_systemd
root 7723 7698 0 00:10 ? 00:00:00 [splunkd pid=7698] splunkd --under-systemd --systemd-delegate=yes -p 8099 _internal_launch_under_systemd [process-runner]
root 7949 5126 0 00:28 pts/0 00:00:00 grep --color=auto splunk
[root@linux_mysql local]#
问题:Splunk_TA_nix add-on下发到Linux Centos7系统上没有采集到cpu性能数据,发现缺少相关命令需要在linux系统上安装以下命令
yum install sysstat
yum install sar
yum insatll mpstat
删除 S P L U N K SPLUNK SPLUNK/etc/instance.cfg (里面记录了guid,也就是客户端名称)
删除 server.conf里面的实例名称记录,inputs.conf里面的主机名称记录
重启 forwarder服务
重新安装SplunkUniversalForwarder
# 例如: 安装目录为/tpdata
/tpdata/splunkforwarder/bin/splunk stop
rm -rf /tpdata/splunkforwarder
tar zxvf /tpdata/splunkforwarder-6*.tgz -C /tpdata
/tpdata/splunkforwarder/bin/splunk start --answer-yes --no-prompt --accept-license
/tpdata/splunkforwarder/bin/splunk set deploy-poll 10.21.8.228:8089 -auth admin:changeme
/tpdata/splunkforwarder/bin/splunk restart
基于splunk clone-prep-clear-config命令清除splunk配置信息创建Linux和Windows脚本
提示:不能修改脚本名称,查找进程是否运行使用Splunk关键字进行查找
linux:clear-Splunkforwarder-config.sh
#!/bin/sh # # Time : 2022/01/12 22:41:30 # Author : Hesc # Version : 1.0 # Splunkforwarder安装路径 Splunkforwarder_DIR="/opt" # Splunkforwarder PID Splunkforwarder_PID=`ps -ef|grep splunk | grep -v grep | awk '{print $2}'` if [ -n "$Splunkforwarder_PID" ]; then echo -e "\033[42;30mINFO: Splunkforwarder guid info:\033[0m" cat $Splunkforwarder_DIR/splunkforwarder/etc/instance.cfg echo -e "\033[41;30mERROR: Splunkforwarder is runing, Stop Splunkforwarder!\033[0m" $Splunkforwarder_DIR/splunkforwarder/bin/splunk stop # 停止splunkforwarder服务 echo -e "\033[42;30mINFO: clear Splunkforwarder config!\033[0m" $Splunkforwarder_DIR/splunkforwarder/bin/splunk clone-prep-clear-config # 清除splunkforwarder配置 echo -e "\033[42;30mINFO: Start Splunkforwarder\033[0m" $Splunkforwarder_DIR/splunkforwarder/bin/splunk start # 启动splunkforwarder服务 echo -e "\033[42;30mINFO: Update Splunkforwarder guid info:\033[0m" cat $Splunkforwarder_DIR/splunkforwarder/etc/instance.cfg else echo -e "\033[42;30mINFO: Splunkforwarder guid info:\033[0m" cat $Splunkforwarder_DIR/splunkforwarder/etc/instance.cfg echo -e "\033[42;30mINFO: Splunkforwarder is not runing, clear Splunkforwarder config!\033[0m" $Splunkforwarder_DIR/splunkforwarder/bin/splunk clone-prep-clear-config # 清除splunkforwarder配置 echo -e "\033[42;30mINFO: Start Splunkforwarder\033[0m" $Splunkforwarder_DIR/splunkforwarder/bin/splunk start # 启动splunkforwarder服务 echo -e "\033[42;30mINFO: Update Splunkforwarder guid info:\033[0m" cat $Splunkforwarder_DIR/splunkforwarder/etc/instance.cfg fi
Windows:clear-Splunkforwarder-config.bat
@echo off :: Time : 2022/01/12 22:41:30 :: Author : Hesc :: Version : 1.0 set SplunkUniversalForwarder_install_dir="C:\Program Files" tasklist|find /i "splunkd.exe" if %errorlevel%==0 ( echo "ERROR: SplunkUniversalForwarder is runing,Stop SplunkUniversalForwarder !!" "C:\Program Files\SplunkUniversalForwarder\bin\splunk.exe" stop "C:\Program Files\SplunkUniversalForwarder\bin\splunk.exe" clone-prep-clear-config "C:\Program Files\SplunkUniversalForwarder\bin\splunk.exe" start ) else ( echo "INFO: SplunkUniversalForwarder is not runing,clear SplunkUniversalForwarder config !!" "C:\Program Files\SplunkUniversalForwarder\bin\splunk.exe" clone-prep-clear-config "C:\Program Files\SplunkUniversalForwarder\bin\splunk.exe" start ) pause
vim /opt/splunk/system/local/limits.conf cat /opt/splunk/system/local/limits.conf [search] max_searches_per_cpu = 1 # 默认值为1 # 每个CPU的最大并发搜索次数
- 1
- 2
- 3
- 4
- 5
- 6
index="_internal" source="/opt/splunk/var/log/splunk/splunkd.log" start
INFO:为正常发送,ERROR:为发送失败 >>要检查ERROR报错的问题
index ="_internal" source="/opt/splunk/var/log/splunk/python.log" "mail"
字段重命名:不能设置多个字段名称重命名为同一个字段名称
重写字段值:如果有设置多个字段名称重命名为同一个字段名称,使用勾选此选项字段重命名。
<row>
<panel> depends="$range$"
<table>
<title>僵尸进程详情</title>
</table>
</panel>
</row>
http://192.168.10.9:8000/zh-CN/debug/refresh
] vim /opt/splunk/splunk/etc/splunk-launch.conf --编辑splunk-launch文件 ] cat /opt/splunk/splunk/etc/splunk-launch.conf # Version 7.2.5 # Modify the following line to suit the location of your Splunk install. # If unset, Splunk will use the parent of the directory containing the splunk # CLI executable. # # SPLUNK_HOME=/opt/splunk-home # By default, Splunk stores its indexes under SPLUNK_HOME in the # var/lib/splunk subdirectory. This can be overridden # here: # SPLUNK_DB=/opt/splunk-home/var/lib/splunk # 添加以下内容 SPLUNK_DB=/opt/splunk/splunk/data # Splunkd daemon name SPLUNK_SERVER_NAME=Splunkd # Splunkweb daemon name SPLUNK_WEB_NAME=splunkweb # If SPLUNK_OS_USER is set, then Splunk service will only start # if the 'splunk [re]start [splunkd]' command is invoked by a user who # is, or can effectively become via setuid(2), $SPLUNK_OS_USER. # (This setting can be specified as username or as UID.) # SPLUNK_OS_USER
] mkdir data
--创建data目录
] splunk stop
--停止splunk服务
] mv /opt/splunk/splunk/var/lib/splunk /opt/splunk/splunk/data/
--将/opt/splunk/splunk/var/lib/splunk目录下的文件移动到/opt/splunk/splunk/data/目录下
] splunk start
--启动服务
$SPLUNK_HOME/etc/licenses/enterprise/Splunk.License.lic
index=wineventlog
| lookup AXA_Server_List Segment as 源网络地址 OUTPUT 分公司区域 机构 省份 Region
| search Region=*
| stats count by 分公司区域 省份 Region 源网络地址 username
| search username!="-" username !="ANONYMOUS LOGON" username!=admin*
| rename 源网络地址 as IP username as nt_host
| rex field=nt_host "(?P<nt_host>[^$]*)"
| eval nt_host=upper(nt_host)
| lookup AXA_ServerIP_List.csv IP nt_host OUTPUT IP as ip
| search NOT ip=*
| fields - ip count
| outputlookup AXA_ServerIP_List.csv append=true
index="_internal" sourcetype="splunkd" source="*metrics.log" fwdType=uf source="/opt/splunk/var/log/splunk/metrics.log"
| stats count by hostname,sourceIp,os,arch,guid
curl -k -u admin:Eccom@123 https://localhost:8089/services/apps/local -d filename="true" -d name="/opt/splunk-enterprise-security_700.spl" -d update="true" -v
如果忘记管理员密码,可以重置。需要有服务器的访问权限。
方法
1.将\$SPLUNK_HOME/etc/passwd文件重命名为passwd.bak
] mv /opt/splunk/etc/passwd passwd.bak
2.重新启动Splunk,此时登录Splunk Web之后,所使用的密码则为初始账号密码:admin changeme
] /opt/splunk/bin/splunk restatr
如果您使用的是Splunk 7.1以上的版本,需要使用<user-seed.conf 创建管理员凭据>方法重置admin密码。
在7.1以前的版本中,缺少$SPLUNK_HOME/etc/passwd
文件, Splunk会将密码重置为"changeme"。但是在7.1上版本此方法已失效。
1. 您需要保存$SPLUNK_HOME/etc/passwd
文件,并删除原始文件。
mv $SPLUNK_HOME/etc/passwd $SPLUNK_HOME/etc/passwd-bak
2.编辑$SPLUNK_HOME/etc/system/local/user-seed.conf
配置文件(不存在则新建)。
添加一个名为[user_info]
章节,并指定管理员用户和您希望使用的密码。
[user_info]
USERNAME=admin
PASSWORD=password
3.然后重新启动Splunk,将会生成一个新的$SPLUNK_HOME/etc/passwd
文件。
splunk restart
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。