当前位置:   article > 正文

YOLOv6 Pro | YOLOv6网络魔改 (1) ——RepGFPN融合高效聚合网络(ELAN)和重参数化的目标检测Neck(来自DAMO-YOLO)

repgfpn

在阿里达摩院 ICLR2022 发表的论文《GiraffeDet: A Heavy-Neck Paradigm for Object Detection》中,他们提出了GiraffeDet,它具有极轻量级计算量的backbone和大计算量的neck,使得网络更关注于高分辨率特征图中空间信息和低分辨率特征图中语义信息的信息交互。同时在2022年11月底他们开源的DAMO YOLO中,再一次用到了GFPN的思想,他们基于queen-fusion的GFPN,加入了高效聚合网络(ELAN)和重参数化的思想,构成了一个新的Neck网络RepGFPN,乘着火热,本篇将在YOLOv6 Pro框架中,在YOLOV6的neck结构中改入RepGFPN,同样改进也可以在YOLOv5中实现。

惯例,介绍下YOLOv6 Pro框架!

 · YOLOv6 Pro 基于官方 YOLOv6 的整体架构,使用 YOLOv5 的网络构建方式构建一个 YOLOv6 网络,包括 backboneneckeffidehead 结构。
· 可以在 yaml 文件中任意修改或添加模块,并且每个修改的文件都是独立可运行的,目的是为了助力科研。
· 后续会基于 yolov5 和 yoloair 中的模块加入更多的网络结构改进。
· 预训练权重已经从官方权重转换,确保可以匹配。

· 预先发布了p6模型(非官方)

· 已经加入了一些其他改进模块,如RepGFPN,FocalTransformer,RepGhost,CoAtNet等

我们使用的 yoloair 和 YOLOv6 pro 框架在 IEEE UV 2022 "Vision Meets Alage" 目标检测竞赛中取得第一名!

项目链接:GitHub - yang-0201/YOLOv6_pro: Make it easier for yolov6 to change the network structure

感兴趣的小伙伴们可以点点Star和Fork,有问题可以及时反馈,项目初期,会对一些功能意见会进行采纳和开发,也欢迎志同道合的朋友来提PR,共同维护和开发项目,项目后续会持续更新和完善,敬请关注!

进入正题!

DAMO YOLO:论文地址 https://arxiv.org/pdf/2211.15444.pdf

这是整个DAMO YOLO的网络结构图,包括基于神经结构搜索NAS技术得到的MAE-NAS主干网络,RepGFPN和一个ZeroHead结构,我们今天主要关注的是RepGFPN结构,可以发现主要的模块是Fusion Block结构。

作者认为GFPN有效的主要原因之一是因为它可以充分交换高级语义信息和低级空间信息。在GFPN中,多尺度特征在前一层和当前层的层次特征中都被融合。更重要的是,log2(n)跳过层连接提供了更有效的信息传输,可以扩展到更深层次的网络。

 同样,当他们在现代yolo系列模型上用GFPN直接替换原先的Neck结构后,可以获得了更高的精度。(后续可以尝试GFPN的替换)但是他们发现的问题在于基于GFPN的模型的延迟远远高于基于改进的panet的模型,所以精度的提升或许有些得不偿失。总结以下几个原因:

1. 不同尺度的特征图具有相同的通道维度;

2. queen-fusion的不能满足实时检测模型的要求;

3. 基于卷积的跨尺度特征融合效率低下;

基于GFPN,我们提出了一种新颖的高效——RepGFPN满足实时目标检测的设计,综合考虑这些原因,作者的思想有以下几点:

1. 由于不同尺度特征图的flop差异较大,在计算成本有限的约束下,很难控制每个尺度特征图共享的相同的通道数。

因此,在作者的的neck特征融合中,采用了不同通道维度的不同尺度特征图的设置。作者比较了相同和不同通道的性能以及neck深度和宽度的权衡,如下表

我们可以看到,通过灵活地控制不同尺度上的信道数量,我们可以获得比在所有尺度上共享相同通道数更高的精度。当深度等于3,宽度等于(96、192、384)时,可以获得最佳的性能。

2) GFPN通过queen-fusion增强了特征交互,但它也带来了大量额外的上采样和降采样操作。

作者比较了这些上采样和下采样操作的性能,结果如表所示

我们可以看到,额外的上采样操作导致延迟增加了0.6 ms,而精度的提高只有0.3mAP,远低于额外的下采样操作带来的性能提升。因此,在实时检测的约束下,我们去掉了queen-fusion中额外的上采样操作。

3. 在特征融合块中,我们首先用CSPNet替换原始的基于3x3卷积的特征融合,获得4.2 mAP增益。之后,我们通过结合重参数化机制和高效层聚合网络(ELAN)的连接来升级CSPNet。由于不带来额外巨大的计算负担,我们实现了更高的精度。比较的结果列于表中

经过论文中的介绍,相信大家已经对 RepGFPN的思想有所了解,下面来看看代码!

我参考DAMO YOLO 的源码在YOLOv6 Pro的框架中加入了RepGFPN结构,包括RepGFPN-T,RepGFPN-M,RepGFPN-S,分别加在了YOLOv6l和YOLOv6t中作为示例

先看yolov6l+RepGFPN-M结构的yaml文件:

  1. depth_multiple: 1.0 # model depth multiple
  2. width_multiple: 1.0 # layer channel multiple
  3. backbone:
  4. # [from, number, module, args]
  5. [[-1, 1, ConvWrapper, [64, 3, 2]], # 0-P1/2
  6. [-1, 1, ConvWrapper, [128, 3, 2]], # 1-P2/4
  7. [-1, 1, BepC3, [128, 6, "ConvWrapper"]],
  8. [-1, 1, ConvWrapper, [256, 3, 2]], # 3-P3/8
  9. [-1, 1, BepC3, [256, 12, "ConvWrapper"]],
  10. [-1, 1, ConvWrapper, [512, 3, 2]], # 5-P4/16
  11. [-1, 1, BepC3, [512, 18, "ConvWrapper"]],
  12. [-1, 1, ConvWrapper, [1024, 3, 2]], # 7-P5/32
  13. [-1, 1, BepC3, [1024, 6, "ConvWrapper"]],
  14. [-1, 1, SPPF, [1024, 5]]] # 9
  15. neck:
  16. [
  17. [ 6, 1,ConvBNAct,[ 256, 3, 2, silu ] ],
  18. [ [ -1, 9 ], 1, Concat, [ 1 ] ], # 768
  19. [ -1, 1, RepGFPN, [ 512, 1.5, 1.0, silu ] ], # 8
  20. [ -1, 1, nn.Upsample, [ None, 2, 'nearest' ] ],
  21. [ 4, 1,ConvBNAct,[ 128, 3, 2, silu ] ],
  22. [ [ -1, 6, 13 ], 1, Concat, [ 1 ] ], # 896
  23. [ -1, 1, RepGFPN, [ 256, 1.5, 1.0, silu ] ], # merge_4 12
  24. [ -1, 1, nn.Upsample, [ None, 2, 'nearest' ] ],
  25. [ [ -1, 4 ], 1, Concat, [ 1 ] ], # 384
  26. [ -1, 1, RepGFPN, [ 128, 1.5, 1.0, silu ] ], # 512+256 merge_5 15 out
  27. [ -1, 1,ConvBNAct,[ 128, 3, 2, silu ] ],
  28. [ [ -1, 16 ], 1, Concat, [ 1 ] ], # 384
  29. [ -1, 1, RepGFPN, [ 256, 1.5, 1.0, silu ] ], # 512+256 merge_7 18 out
  30. [ 16, 1,ConvBNAct,[ 256, 3, 2, silu ] ],
  31. [ -2, 1,ConvBNAct,[ 256, 3, 2, silu ] ],
  32. [ [ -1, 12, -2 ], 1, Concat, [ 1 ] ], # 1024
  33. [ -1, 1, RepGFPN, [ 512, 1.5, 1.0, silu ] ], # 512+512+1024 merge_6 22 out
  34. ]
  35. effidehead:
  36. [[19, 1,Head_out , [128, 16]],
  37. [22, 1, Head_out, [256, 16]],
  38. [26, 1, Head_out, [512, 16]],
  39. [[27, 28, 29], 1, Out, []]]

 对比下原图

其中各种颜色的模块就是ConvBNAct,是一个简单的卷积加标准化加relu/silu激活函数的模块。

Fusion Block就是主要的融合模块,对应yaml文件中的RepGFPN

 输入为两个或者三个层,经过concat之后,分别用1x1卷积降通道,下方就是模仿ELAN的特征聚合模块,由N个Rep 3x3卷积和3x3卷积组成,不同层同时输出,再通过concat得到最终的输出。

 输入RepGFPN的参数含义为[输出通道,深度系数,中间层通道的缩放因子,使用的激活函数类型],为了保持重Neck,轻Head的思想,我去掉了6的解耦头,换为了Head_out 做一个简单输出

yolov6t+RepGFPN-T结构的yaml文件:

  1. depth_multiple: 0.33 # model depth multiple
  2. width_multiple: 0.375 # layer channel multiple
  3. backbone:
  4. # [from, number, module, args]
  5. [[-1, 1, RepVGGBlock, [64, 3, 2]], # 0-P1/2
  6. [-1, 1, RepVGGBlock, [128, 3, 2]], # 1-P2/4
  7. [-1, 6, RepBlock, [128]],
  8. [-1, 1, RepVGGBlock, [256, 3, 2]], # 3-P3/8
  9. [-1, 12, RepBlock, [256]],
  10. [-1, 1, RepVGGBlock, [512, 3, 2]], # 5-P4/16
  11. [-1, 18, RepBlock, [512]],
  12. [-1, 1, RepVGGBlock, [1024, 3, 2]], # 7-P5/32
  13. [-1, 6, RepBlock, [1024]],
  14. [-1, 1, SimSPPF, [1024, 5]]] # 9
  15. neck:
  16. [
  17. [ 6, 1,ConvBNAct,[ 192, 3, 2 ] ],
  18. [ [ -1, 9 ], 1, Concat, [ 1 ] ], # 576
  19. [ -1, 1, RepGFPN, [ 384, 1.0, 1.0 ] ], # 8
  20. [ -1, 1, nn.Upsample, [ None, 2, 'nearest' ] ],
  21. [ 4, 1,ConvBNAct,[ 96, 3, 2 ] ],
  22. [ [ -1, 6, 13 ], 1, Concat, [ 1 ] ], # 672
  23. [ -1, 1, RepGFPN, [ 192, 1.0, 1.0 ] ], # merge_4 12
  24. [ -1, 1, nn.Upsample, [ None, 2, 'nearest' ] ],
  25. [ [ -1, 4 ], 1, Concat, [ 1 ] ], # 288
  26. [ -1, 1, RepGFPN, [ 64, 1.0, 1.0 ] ], # merge_5 15 out
  27. [ -1, 1,ConvBNAct,[ 64, 3, 2 ] ],
  28. [ [ -1, 16 ], 1, Concat, [ 1 ] ], # 256
  29. [ -1, 1, RepGFPN, [ 128, 1.0, 1.0 ] ], # merge_7 18 out
  30. [ 16, 1,ConvBNAct,[ 192, 3, 2 ] ],
  31. [ -2, 1,ConvBNAct,[ 128, 3, 2 ] ],
  32. [ [ -1, 12, -2 ], 1, Concat, [ 1 ] ], # 704
  33. [ -1, 1, RepGFPN, [ 256, 1.0, 1.0 ] ], # merge_6 22 out
  34. ]
  35. effidehead:
  36. [[19, 1,Head_out , [170, 0]], ##170 * 0.375 = 64
  37. [22, 1, Head_out, [341, 0]], ##341 * 0.375 = 128
  38. [26, 1, Head_out, [682, 0]], ##682 * 0.375 = 256
  39. [[27, 28, 29], 1, Out, []]]

需要加入的代码为,在common.py中加入,或者自己新建一个RepGFPN.py文件,再在common.py中导入模块名称,

from yolov6.layers.damo_yolo import ConvBNAct,RepGFPN
  1. import numpy as np
  2. import torch
  3. import torch.nn as nn
  4. class RepGFPN(nn.Module):
  5. def __init__(self,in_channels,out_channels,depth=1.0,hidden_ratio = 1.0,act = 'relu',block_name='BasicBlock_3x3_Reverse',spp = False):
  6. super(RepGFPN, self).__init__()
  7. self.merge_3 = CSPStage(block_name,
  8. in_channels,
  9. hidden_ratio,
  10. out_channels,
  11. round(3 * depth),
  12. act=act)
  13. def forward(self,x):
  14. x = self.merge_3(x)
  15. return x
  16. class CSPStage(nn.Module):
  17. def __init__(self,
  18. block_fn,
  19. ch_in,
  20. ch_hidden_ratio,
  21. ch_out,
  22. n,
  23. act='swish',
  24. spp=False):
  25. super(CSPStage, self).__init__()
  26. split_ratio = 2
  27. ch_first = int(ch_out // split_ratio)
  28. ch_mid = int(ch_out - ch_first)
  29. self.conv1 = ConvBNAct(ch_in, ch_first, 1, act=act)
  30. self.conv2 = ConvBNAct(ch_in, ch_mid, 1, act=act)
  31. self.convs = nn.Sequential()
  32. next_ch_in = ch_mid
  33. for i in range(n):
  34. if block_fn == 'BasicBlock_3x3_Reverse':
  35. self.convs.add_module(
  36. str(i),
  37. BasicBlock_3x3_Reverse(next_ch_in,
  38. ch_hidden_ratio,
  39. ch_mid,
  40. act=act,
  41. shortcut=True))
  42. else:
  43. raise NotImplementedError
  44. if i == (n - 1) // 2 and spp:
  45. self.convs.add_module(
  46. 'spp', SPP(ch_mid * 4, ch_mid, 1, [5, 9, 13], act=act))
  47. next_ch_in = ch_mid
  48. self.conv3 = ConvBNAct(ch_mid * n + ch_first, ch_out, 1, act=act)
  49. def forward(self, x):
  50. y1 = self.conv1(x)
  51. y2 = self.conv2(x)
  52. mid_out = [y1]
  53. for conv in self.convs:
  54. y2 = conv(y2)
  55. mid_out.append(y2)
  56. y = torch.cat(mid_out, axis=1)
  57. y = self.conv3(y)
  58. return y
  59. class ConvBNAct(nn.Module):
  60. """A Conv2d -> Batchnorm -> silu/leaky relu block"""
  61. def __init__(
  62. self,
  63. in_channels,
  64. out_channels,
  65. ksize,
  66. stride=1,
  67. act='relu',
  68. groups=1,
  69. bias=False,
  70. norm='bn',
  71. reparam=False,
  72. ):
  73. super().__init__()
  74. # same padding
  75. pad = (ksize - 1) // 2
  76. self.conv = nn.Conv2d(
  77. in_channels,
  78. out_channels,
  79. kernel_size=ksize,
  80. stride=stride,
  81. padding=pad,
  82. groups=groups,
  83. bias=bias,
  84. )
  85. if norm is not None:
  86. self.bn = get_norm(norm, out_channels, inplace=True)
  87. if act is not None:
  88. self.act = get_activation(act, inplace=True)
  89. self.with_norm = norm is not None
  90. self.with_act = act is not None
  91. def forward(self, x):
  92. x = self.conv(x)
  93. if self.with_norm:
  94. x = self.bn(x)
  95. if self.with_act:
  96. x = self.act(x)
  97. return x
  98. def fuseforward(self, x):
  99. return self.act(self.conv(x))
  100. class BasicBlock_3x3_Reverse(nn.Module):
  101. def __init__(self,
  102. ch_in,
  103. ch_hidden_ratio,
  104. ch_out,
  105. act='relu',
  106. shortcut=True):
  107. super(BasicBlock_3x3_Reverse, self).__init__()
  108. assert ch_in == ch_out
  109. ch_hidden = int(ch_in * ch_hidden_ratio)
  110. self.conv1 = ConvBNAct(ch_hidden, ch_out, 3, stride=1, act=act)
  111. self.conv2 = RepConv(ch_in, ch_hidden, 3, stride=1, act=act)
  112. self.shortcut = shortcut
  113. def forward(self, x):
  114. y = self.conv2(x)
  115. y = self.conv1(y)
  116. if self.shortcut:
  117. return x + y
  118. else:
  119. return y
  120. def get_norm(name, out_channels, inplace=True):
  121. if name == 'bn':
  122. module = nn.BatchNorm2d(out_channels)
  123. else:
  124. raise NotImplementedError
  125. return module
  126. class SPP(nn.Module):
  127. def __init__(
  128. self,
  129. ch_in,
  130. ch_out,
  131. k,
  132. pool_size,
  133. act='swish',
  134. ):
  135. super(SPP, self).__init__()
  136. self.pool = []
  137. for i, size in enumerate(pool_size):
  138. pool = nn.MaxPool2d(kernel_size=size,
  139. stride=1,
  140. padding=size // 2,
  141. ceil_mode=False)
  142. self.add_module('pool{}'.format(i), pool)
  143. self.pool.append(pool)
  144. self.conv = ConvBNAct(ch_in, ch_out, k, act=act)
  145. def forward(self, x):
  146. outs = [x]
  147. for pool in self.pool:
  148. outs.append(pool(x))
  149. y = torch.cat(outs, axis=1)
  150. y = self.conv(y)
  151. return y
  152. import torch.nn.functional as F
  153. class Swish(nn.Module):
  154. def __init__(self, inplace=True):
  155. super(Swish, self).__init__()
  156. self.inplace = inplace
  157. def forward(self, x):
  158. if self.inplace:
  159. x.mul_(F.sigmoid(x))
  160. return x
  161. else:
  162. return x * F.sigmoid(x)
  163. class RepConv(nn.Module):
  164. '''RepConv is a basic rep-style block, including training and deploy status
  165. Code is based on https://github.com/DingXiaoH/RepVGG/blob/main/repvgg.py
  166. '''
  167. def __init__(self,
  168. in_channels,
  169. out_channels,
  170. kernel_size=3,
  171. stride=1,
  172. padding=1,
  173. dilation=1,
  174. groups=1,
  175. padding_mode='zeros',
  176. deploy=False,
  177. act='relu',
  178. norm=None):
  179. super(RepConv, self).__init__()
  180. self.deploy = deploy
  181. self.groups = groups
  182. self.in_channels = in_channels
  183. self.out_channels = out_channels
  184. assert kernel_size == 3
  185. assert padding == 1
  186. padding_11 = padding - kernel_size // 2
  187. if isinstance(act, str):
  188. self.nonlinearity = get_activation(act)
  189. else:
  190. self.nonlinearity = act
  191. if deploy:
  192. self.rbr_reparam = nn.Conv2d(in_channels=in_channels,
  193. out_channels=out_channels,
  194. kernel_size=kernel_size,
  195. stride=stride,
  196. padding=padding,
  197. dilation=dilation,
  198. groups=groups,
  199. bias=True,
  200. padding_mode=padding_mode)
  201. else:
  202. self.rbr_identity = None
  203. self.rbr_dense = conv_bn(in_channels=in_channels,
  204. out_channels=out_channels,
  205. kernel_size=kernel_size,
  206. stride=stride,
  207. padding=padding,
  208. groups=groups)
  209. self.rbr_1x1 = conv_bn(in_channels=in_channels,
  210. out_channels=out_channels,
  211. kernel_size=1,
  212. stride=stride,
  213. padding=padding_11,
  214. groups=groups)
  215. def forward(self, inputs):
  216. '''Forward process'''
  217. if hasattr(self, 'rbr_reparam'):
  218. return self.nonlinearity(self.rbr_reparam(inputs))
  219. if self.rbr_identity is None:
  220. id_out = 0
  221. else:
  222. id_out = self.rbr_identity(inputs)
  223. return self.nonlinearity(
  224. self.rbr_dense(inputs) + self.rbr_1x1(inputs) + id_out)
  225. def get_equivalent_kernel_bias(self):
  226. kernel3x3, bias3x3 = self._fuse_bn_tensor(self.rbr_dense)
  227. kernel1x1, bias1x1 = self._fuse_bn_tensor(self.rbr_1x1)
  228. kernelid, biasid = self._fuse_bn_tensor(self.rbr_identity)
  229. return kernel3x3 + self._pad_1x1_to_3x3_tensor(
  230. kernel1x1) + kernelid, bias3x3 + bias1x1 + biasid
  231. def _pad_1x1_to_3x3_tensor(self, kernel1x1):
  232. if kernel1x1 is None:
  233. return 0
  234. else:
  235. return torch.nn.functional.pad(kernel1x1, [1, 1, 1, 1])
  236. def _fuse_bn_tensor(self, branch):
  237. if branch is None:
  238. return 0, 0
  239. if isinstance(branch, nn.Sequential):
  240. kernel = branch.conv.weight
  241. running_mean = branch.bn.running_mean
  242. running_var = branch.bn.running_var
  243. gamma = branch.bn.weight
  244. beta = branch.bn.bias
  245. eps = branch.bn.eps
  246. else:
  247. assert isinstance(branch, nn.BatchNorm2d)
  248. if not hasattr(self, 'id_tensor'):
  249. input_dim = self.in_channels // self.groups
  250. kernel_value = np.zeros((self.in_channels, input_dim, 3, 3),
  251. dtype=np.float32)
  252. for i in range(self.in_channels):
  253. kernel_value[i, i % input_dim, 1, 1] = 1
  254. self.id_tensor = torch.from_numpy(kernel_value).to(
  255. branch.weight.device)
  256. kernel = self.id_tensor
  257. running_mean = branch.running_mean
  258. running_var = branch.running_var
  259. gamma = branch.weight
  260. beta = branch.bias
  261. eps = branch.eps
  262. std = (running_var + eps).sqrt()
  263. t = (gamma / std).reshape(-1, 1, 1, 1)
  264. return kernel * t, beta - running_mean * gamma / std
  265. def switch_to_deploy(self):
  266. if hasattr(self, 'rbr_reparam'):
  267. return
  268. kernel, bias = self.get_equivalent_kernel_bias()
  269. self.rbr_reparam = nn.Conv2d(
  270. in_channels=self.rbr_dense.conv.in_channels,
  271. out_channels=self.rbr_dense.conv.out_channels,
  272. kernel_size=self.rbr_dense.conv.kernel_size,
  273. stride=self.rbr_dense.conv.stride,
  274. padding=self.rbr_dense.conv.padding,
  275. dilation=self.rbr_dense.conv.dilation,
  276. groups=self.rbr_dense.conv.groups,
  277. bias=True)
  278. self.rbr_reparam.weight.data = kernel
  279. self.rbr_reparam.bias.data = bias
  280. for para in self.parameters():
  281. para.detach_()
  282. self.__delattr__('rbr_dense')
  283. self.__delattr__('rbr_1x1')
  284. if hasattr(self, 'rbr_identity'):
  285. self.__delattr__('rbr_identity')
  286. if hasattr(self, 'id_tensor'):
  287. self.__delattr__('id_tensor')
  288. self.deploy = True
  289. def get_activation(name='silu', inplace=True):
  290. if name is None:
  291. return nn.Identity()
  292. if isinstance(name, str):
  293. if name == 'silu':
  294. module = nn.SiLU(inplace=inplace)
  295. elif name == 'relu':
  296. module = nn.ReLU(inplace=inplace)
  297. elif name == 'lrelu':
  298. module = nn.LeakyReLU(0.1, inplace=inplace)
  299. elif name == 'swish':
  300. module = Swish(inplace=inplace)
  301. elif name == 'hardsigmoid':
  302. module = nn.Hardsigmoid(inplace=inplace)
  303. elif name == 'identity':
  304. module = nn.Identity()
  305. else:
  306. raise AttributeError('Unsupported act type: {}'.format(name))
  307. return module
  308. elif isinstance(name, nn.Module):
  309. return name
  310. else:
  311. raise AttributeError('Unsupported act type: {}'.format(name))
  312. def conv_bn(in_channels, out_channels, kernel_size, stride, padding, groups=1):
  313. '''Basic cell for rep-style block, including conv and bn'''
  314. result = nn.Sequential()
  315. result.add_module(
  316. 'conv',
  317. nn.Conv2d(in_channels=in_channels,
  318. out_channels=out_channels,
  319. kernel_size=kernel_size,
  320. stride=stride,
  321. padding=padding,
  322. groups=groups,
  323. bias=False))
  324. result.add_module('bn', nn.BatchNorm2d(num_features=out_channels))
  325. return result

在yolo.py 中加入

  1. elif m in [RepGFPN, ConvBnAct]:
  2. c1 = ch[f]
  3. c2 = args[0]
  4. args = [c1, c2, *args[1:]]

大功告成!

声明:本文内容由网友自发贡献,不代表【wpsshop博客】立场,版权归原作者所有,本站不承担相应法律责任。如您发现有侵权的内容,请联系我们。转载请注明出处:https://www.wpsshop.cn/w/Cpp五条/article/detail/371365
推荐阅读
相关标签
  

闽ICP备14008679号