赞
踩
看一下项目的目录结构
minGPT 库由三个文件组成
minGPT/model.py
包含实际的 Transformer 模型定义minGPT/bpe.py
包含一个稍微重构的 Byte Pair 编码器,它用于将文本转换为整数序列,为 tokenize 做准备minGPT/trainer.py
是训练模型的 (与GPT无关的) PyTorch样板代码在 projects 文件夹中给出了一些使用 minGPT 库的 demos 和 projects,包括
projects/adder
从零训练 GPT 做加法(灵感来自GPT-3论文中的添加部分)projects/chargpt
用一些输入文本将 GPT 训练成一个 character-level language modeldemo.ipynb
在一个简单的排序示例中以笔记本格式展示了 GPT
和 Trainer
的 minimal usagegenerate.ipynb
展示了如何加载预训练的 GPT2 并在给定 prompt 的情况下生成文本minGPT/model.py
中的 GPT 模型实现先看下 Transformer block 的定义,它是 GPT 的核心组件
class Block(nn.Module): """ an unassuming Transformer block """ def __init__(self, config): super().__init__() self.ln_1 = nn.LayerNorm(config.n_embd) self.attn = CausalSelfAttention(config) self.ln_2 = nn.LayerNorm(config.n_embd) self.mlp = nn.ModuleDict(dict( c_fc = nn.Linear(config.n_embd, 4 * config.n_embd), c_proj = nn.Linear(4 * config.n_embd, config.n_embd), act = NewGELU(), dropout = nn.Dropout(config.resid_pdrop), )) m = self.mlp self.mlpf = lambda x: m.dropout(m.c_proj(m.act(m.c_fc(x)))) # MLP forward def forward(self, x): x = x + self.attn(self.ln_1(x)) x = x + self.mlpf(self.ln_2(x)) return x
结构很清晰,如下图所示
可见使用了 GPT2/3 的 LayerNorm 前置结构,其中两个核心模块是 masked self-attention 层 self.attn
和 FFD 层 self.mlp
FFD 层:很简单就是一个带 dropout 正则的两层 MLP,特殊之处在于使用了 gelu 激活函数,如下所示
class NewGELU(nn.Module):
"""
Implementation of the GELU activation function currently in Google BERT repo (identical to OpenAI GPT).
Reference: Gaussian Error Linear Units (GELU) paper: https://arxiv.org/abs/1606.08415
"""
def forward(self, x):
return 0.5 * x * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (x + 0.044715 * torch.pow(x, 3.0))))
这个激活函数是对 relu 的一个优化,增加了更多非线性成分,如下所示
masked self attention 层:这是 GPT 模型的核心,需要计算因果注意力
class CausalSelfAttention(nn.Module): """ A vanilla multi-head masked self-attention layer with a projection at the end. It is possible to use torch.nn.MultiheadAttention here but I am including an explicit implementation here to show that there is nothing too scary here. """ def __init__(self, config): super().__init__() assert config.n_embd % config.n_head == 0 # key, query, value projections for all heads, but in a batch self.c_attn = nn.Linear(config.n_embd, 3 * config.n_embd) # output projection self.c_proj = nn.Linear(config.n_embd, config.n_embd) # regularization self.attn_dropout = nn.Dropout(config.attn_pdrop) self.resid_dropout = nn.Dropout(config.resid_pdrop) # causal mask to ensure that attention is only applied to the left in the input sequence self.register_buffer("bias", torch.tril(torch.ones(config.block_size, config.block_size)) .view(1, 1, config.block_size, config.block_size)) self.n_head = config.n_head self.n_embd = config.n_embd def forward(self, x): B, T, C = x.size() # batch size, sequence length, embedding dimensionality (n_embd) # calculate query, key, values for all heads in batch and move head forward to be the batch dim q, k ,v = self.c_attn(x).split(self.n_embd, dim=2) k = k.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) q = q.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) v = v.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs) # causal self-attention; Self-attend: (B, nh, T, hs) x (B, nh, hs, T) -> (B, nh, T, T) att = (q @ k.transpose(-2, -1)) * (1.0 / math.sqrt(k.size(-1))) att = att.masked_fill(self.bias[:,:,:T,:T] == 0, float('-inf')) att = F.softmax(att, dim=-1) att = self.attn_dropout(att) y = att @ v # (B, nh, T, T) x (B, nh, T, hs) -> (B, nh, T, hs) y = y.transpose(1, 2).contiguous().view(B, T, C) # re-assemble all head outputs side by side # output projection y = self.resid_dropout(self.c_proj(y)) return y
hs
为嵌入维度 n_embd
的 1/n_head。虽然编程时这里仅需保证 query 和 key 的空间维度一致即可,但是考虑到嵌入维度和 vocab table 大小的关系,每个注意力头处理的嵌入维度不宜过大,这样和 n_embd 绑定是比较方便的做法。请参考:最小熵原理(六):词向量的维度应该怎么选择?self.bias
来实现,序列长度(config.block_size
)为 5 时该张量形如tensor([[[[1., 0., 0., 0., 0.],
[1., 1., 0., 0., 0.],
[1., 1., 1., 0., 0.],
[1., 1., 1., 1., 0.],
[1., 1., 1., 1., 1.]]]])
att.masked_fill(self.bias[:,:,:T,:T] == 0, float('-inf'))
这句就会把这些为 0 的位置替换为 -inf,从而使这些位置在经过 softmax 后权重趋近 0,实现对未来序列的遮盖。另需注意 bias
张量是通过 self.register_buffer
方法登记的,这样登记过的张量可以求梯度也可以随模型在 CPU/GPU 之间移动,但是不进行参数优化class GPT(nn.Module): """ GPT Language Model """ @staticmethod def get_default_config(): C = CfgNode() # either model_type or (n_layer, n_head, n_embd) must be given in the config C.model_type = 'gpt' C.n_layer = None C.n_head = None C.n_embd = None # these options must be filled in externally C.vocab_size = None C.block_size = None # dropout hyperparameters C.embd_pdrop = 0.1 C.resid_pdrop = 0.1 C.attn_pdrop = 0.1 return C def __init__(self, config): super().__init__() assert config.vocab_size is not None assert config.block_size is not None self.block_size = config.block_size type_given = config.model_type is not None params_given = all([config.n_layer is not None, config.n_head is not None, config.n_embd is not None]) assert type_given ^ params_given # exactly one of these (XOR) if type_given: # translate from model_type to detailed configuration config.merge_from_dict({ # names follow the huggingface naming conventions # GPT-1 'openai-gpt': dict(n_layer=12, n_head=12, n_embd=768), # 117M params # GPT-2 configs 'gpt2': dict(n_layer=12, n_head=12, n_embd=768), # 124M params 'gpt2-medium': dict(n_layer=24, n_head=16, n_embd=1024), # 350M params 'gpt2-large': dict(n_layer=36, n_head=20, n_embd=1280), # 774M params 'gpt2-xl': dict(n_layer=48, n_head=25, n_embd=1600), # 1558M params # Gophers 'gopher-44m': dict(n_layer=8, n_head=16, n_embd=512), # (there are a number more...) # I made these tiny models up 'gpt-mini': dict(n_layer=6, n_head=6, n_embd=192), 'gpt-micro': dict(n_layer=4, n_head=4, n_embd=128), 'gpt-nano': dict(n_layer=3, n_head=3, n_embd=48), }[config.model_type]) self.transformer = nn.ModuleDict(dict( wte = nn.Embedding(config.vocab_size, config.n_embd), wpe = nn.Embedding(config.block_size, config.n_embd), drop = nn.Dropout(config.embd_pdrop), h = nn.ModuleList([Block(config) for _ in range(config.n_layer)]), ln_f = nn.LayerNorm(config.n_embd), )) self.lm_head = nn.Linear(config.n_embd, config.vocab_size, bias=False) # init all weights, and apply a special scaled init to the residual projections, per GPT-2 paper self.apply(self._init_weights) for pn, p in self.named_parameters(): if pn.endswith('c_proj.weight'): torch.nn.init.normal_(p, mean=0.0, std=0.02/math.sqrt(2 * config.n_layer)) # report number of parameters (note we don't count the decoder parameters in lm_head) n_params = sum(p.numel() for p in self.transformer.parameters()) print("number of parameters: %.2fM" % (n_params/1e6,)) def _init_weights(self, module): if isinstance(module, nn.Linear): torch.nn.init.normal_(module.weight, mean=0.0, std=0.02) if module.bias is not None: torch.nn.init.zeros_(module.bias) elif isinstance(module, nn.Embedding): torch.nn.init.normal_(module.weight, mean=0.0, std=0.02) elif isinstance(module, nn.LayerNorm): torch.nn.init.zeros_(module.bias) torch.nn.init.ones_(module.weight)
CfgNode
对象来管理参数,这个是一个类似 yacs 的轻量级参数管理类对象,具体实现在 minGPT/utils.py
文件中,请自行查阅。另外注意到作者设置了一系列预制参数,可以轻松定义不同规模的 GPT 模型self.transformer
成员中,wte
和 wpe
分别是 token embedding 和 position embedding 模块,注意到它们都是 nn.Embedding
层;’h
中包含了所有堆叠的 transformer block 层;另外 self.lm_head
是 GPT 模型最后的分类头,GPT 模型宏观上就是一个输出空间为 vocab_size
的分类器self.apply(self._init_weights)
对模型参数进行了初始化,这一通用性是比较强的,可以记录下复用到自己的项目中@classmethod def from_pretrained(cls, model_type): """ Initialize a pretrained GPT model by copying over the weights from a huggingface/transformers checkpoint. """ assert model_type in {'gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'} from transformers import GPT2LMHeadModel # create a from-scratch initialized minGPT model config = cls.get_default_config() config.model_type = model_type config.vocab_size = 50257 # openai's model vocabulary config.block_size = 1024 # openai's model block_size model = GPT(config) sd = model.state_dict() # init a huggingface/transformers model model_hf = GPT2LMHeadModel.from_pretrained(model_type) sd_hf = model_hf.state_dict() # copy while ensuring all of the parameters are aligned and match in names and shapes keys = [k for k in sd_hf if not k.endswith('attn.masked_bias')] # ignore these transposed = ['attn.c_attn.weight', 'attn.c_proj.weight', 'mlp.c_fc.weight', 'mlp.c_proj.weight'] # basically the openai checkpoints use a "Conv1D" module, but we only want to use a vanilla nn.Linear. # this means that we have to transpose these weights when we import them assert len(keys) == len(sd) for k in keys: if any(k.endswith(w) for w in transposed): # special treatment for the Conv1D weights we need to transpose assert sd_hf[k].shape[::-1] == sd[k].shape with torch.no_grad(): sd[k].copy_(sd_hf[k].t()) else: # vanilla copy over the other parameters assert sd_hf[k].shape == sd[k].shape with torch.no_grad(): sd[k].copy_(sd_hf[k]) return model
def configure_optimizers(self, train_config): """ This long function is unfortunately doing something very simple and is being very defensive: We are separating out all parameters of the model into two buckets: those that will experience weight decay for regularization and those that won't (biases, and layernorm/embedding weights). We are then returning the PyTorch optimizer object. """ # separate out all parameters to those that will and won't experience regularizing weight decay decay = set() no_decay = set() whitelist_weight_modules = (torch.nn.Linear, ) blacklist_weight_modules = (torch.nn.LayerNorm, torch.nn.Embedding) for mn, m in self.named_modules(): for pn, p in m.named_parameters(): fpn = '%s.%s' % (mn, pn) if mn else pn # full param name # random note: because named_modules and named_parameters are recursive # we will see the same tensors p many many times. but doing it this way # allows us to know which parent module any tensor p belongs to... if pn.endswith('bias'): # all biases will not be decayed no_decay.add(fpn) elif pn.endswith('weight') and isinstance(m, whitelist_weight_modules): # weights of whitelist modules will be weight decayed decay.add(fpn) elif pn.endswith('weight') and isinstance(m, blacklist_weight_modules): # weights of blacklist modules will NOT be weight decayed no_decay.add(fpn) # validate that we considered every parameter param_dict = {pn: p for pn, p in self.named_parameters()} inter_params = decay & no_decay union_params = decay | no_decay assert len(inter_params) == 0, "parameters %s made it into both decay/no_decay sets!" % (str(inter_params), ) assert len(param_dict.keys() - union_params) == 0, "parameters %s were not separated into either decay/no_decay set!" \ % (str(param_dict.keys() - union_params), ) # create the pytorch optimizer object optim_groups = [ {"params": [param_dict[pn] for pn in sorted(list(decay))], "weight_decay": train_config.weight_decay}, {"params": [param_dict[pn] for pn in sorted(list(no_decay))], "weight_decay": 0.0}, ] optimizer = torch.optim.AdamW(optim_groups, lr=train_config.learning_rate, betas=train_config.betas) return optimizer
torch.nn.Linear
层的 weight
参数进行衰减,bias
参数及所有 torch.nn.LayerNorm
、torch.nn.Embedding
模块的参数都不做处理。由于模块是递归组织的,这个二重变量会重复访问很多参数,所以通过 set
自动去重,最后根据处理结果定义 torch.optim.AdamW
优化器返回
关于权重衰减的理论说明,参考:机器学习基础(6)—— 使用权重衰减和丢弃法缓解过拟合问题
def forward(self, idx, targets=None): device = idx.device b, t = idx.size() assert t <= self.block_size, f"Cannot forward sequence of length {t}, block size is only {self.block_size}" pos = torch.arange(0, t, dtype=torch.long, device=device).unsqueeze(0) # shape (1, t) # forward the GPT model itself tok_emb = self.transformer.wte(idx) # (b, t, n_embd) pos_emb = self.transformer.wpe(pos) # (1, t, n_embd) x = self.transformer.drop(tok_emb + pos_emb) # (b, t, n_embd) for block in self.transformer.h: x = block(x) x = self.transformer.ln_f(x) # (b, t, n_embd) logits = self.lm_head(x) # (b, t, vocab_size) # if we are given some desired targets also calculate the loss loss = None if targets is not None: loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1), ignore_index=-1) return logits, loss
idx
,用 wte
进行 embedding 并加上 wpe
位置编码后得到尺寸 (batch_size, t, n_embd) 的 token embedding 序列 x
,经过若干 transformer block 和后置 layer norm 层 ln_f
,最后经过分类头 self.lm_head
得到尺寸 (batch_size, t, n_embd) 的预测序列 logits
logits
序列长度为 ttargets
序列。将 x
和 targets
都拉平来计算分类的交叉熵损失。注意交叉熵函数中设置了 ignore_index=-1
,这样我们可以对 target 序列设置一些 -1 来 mask 掉无需计算 loss 的 prompt 序列@torch.no_grad() def generate(self, idx, max_new_tokens, temperature=1.0, do_sample=False, top_k=None): """ Take a conditioning sequence of indices idx (LongTensor of shape (b,t)) and complete the sequence max_new_tokens times, feeding the predictions back into the model each time. Most likely you'll want to make sure to be in model.eval() mode of operation for this. """ for _ in range(max_new_tokens): # if the sequence context is growing too long we must crop it at block_size idx_cond = idx if idx.size(1) <= self.block_size else idx[:, -self.block_size:] # (batch_size, seq_len) # forward the model to get the logits for the index in the sequence logits, _ = self(idx_cond) # (batch_size, seq_len, vocab_size) # pluck the logits at the final step and scale by desired temperature logits = logits[:, -1, :] / temperature # (batch_size, vocab_size) # optionally crop the logits to only the top k options if top_k is not None: v, _ = torch.topk(logits, top_k) logits[logits < v[:, [-1]]] = -float('Inf') # apply softmax to convert logits to (normalized) probabilities probs = F.softmax(logits, dim=-1) # (batch_size, vocab_size) # either sample from the distribution or take the most likely element if do_sample: idx_next = torch.multinomial(probs, num_samples=1) # (batch_size, 1) else: _, idx_next = torch.topk(probs, k=1, dim=-1) # append sampled index to the running sequence and continue idx = torch.cat((idx, idx_next), dim=1) # (batch_size, seq_len+1) return idx
idx
,这时序列长度 seq_len 完全是 prompt 长度logits
logits = logits[:, -1, :] / temperature
拿出最后一个位置的预测idx_next
,将其拼接回原序列projects/adder.py
中的两位数加法实验本实验仅考虑两个不超过两位的整数间的加法,这样计算的结果可能是两位数或者三位数。下面给出两个加法算式以及其对应的序列样本
85
+
50
=
135
⟶
8550531
6
+
39
=
45
⟶
0639054
85 + 50 = 135 \longrightarrow 8550531 \\ 6 + 39 = 45 \longrightarrow 0639054
85+50=135⟶85505316+39=45⟶0639054
下面给出数据集的构造代码
class AdditionDataset(Dataset): """ Creates n-digit addition problems. For example, if n=2, then an example addition problem would be to add 85 + 50 = 135. This problem would be represented as the following string for the GPT: "8550531" This is because: - we are discarding the + and =, which are not necessary. We just encode the digits of the input numbers concatenated together. - the result 135 is encoded backwards to make the addition easier to learn for the GPT model, because of how the addition algorithm works. As one more example, the problem 6 + 39 = 45 would be encoded as: "0639054" where you will notice that we are padding with zeros to make sure that we always produce strings of the exact same size: n + n + (n + 1). When n=2, this is 7. At test time, we will feed in an addition problem by giving the first 2n digits, and hoping that the GPT model completes the sequence with the next (n+1) digits correctly. """ @staticmethod def get_default_config(): C = CN() C.ndigit = 2 return C def __init__(self, config, split): self.config = config self.split = split # train/test # split up all addition problems into either training data or test data ndigit = self.config.ndigit assert ndigit <= 3, "the lines below would be very memory inefficient, in future maybe refactor to support" num = (10**ndigit)**2 # total number of possible addition problems with ndigit numbers rng = torch.Generator() rng.manual_seed(1337) perm = torch.randperm(num, generator=rng) num_test = min(int(num*0.2), 500) # 20% of the whole dataset, or only up to 500 self.ixes = perm[:num_test] if split == 'test' else perm[num_test:] def get_vocab_size(self): return 10 # digits 0..9 def get_block_size(self): # a,b,a+b, and +1 due to potential carry overflow, # but then also -1 because very last digit doesn't ever plug back # as there is no explicit <EOS> token to predict, it is implied return 3*self.config.ndigit + 1 - 1 def __len__(self): return self.ixes.nelement() def __getitem__(self, idx): ndigit = self.config.ndigit # given a problem index idx, first recover the associated a + b idx = self.ixes[idx].item() nd = 10**ndigit a = idx // nd b = idx % nd # calculate the "label" of the addition problem a + b c = a + b # encode the digits of a, b, c into strings astr = f'%0{ndigit}d' % a bstr = f'%0{ndigit}d' % b cstr = (f'%0{ndigit+1}d' % c)[::-1] # reverse c to make addition easier render = astr + bstr + cstr dix = [int(s) for s in render] # convert each character to its token index # x will be input to GPT and y will be the associated expected outputs x = torch.tensor(dix[:-1], dtype=torch.long) y = torch.tensor(dix[1:], dtype=torch.long) # predict the next token in the sequence y[:ndigit*2-1] = -1 # we will only train in the output locations. -1 will mask loss to zero return x, y
打印一些生成的样本和标签看一看
(tensor([9, 9, 3, 2, 1, 3]), tensor([-1, -1, -1, 1, 3, 1]))
(tensor([5, 6, 2, 1, 7, 7]), tensor([-1, -1, -1, 7, 7, 0]))
(tensor([5, 5, 0, 3, 8, 5]), tensor([-1, -1, -1, 8, 5, 0]))
(tensor([3, 8, 6, 2, 0, 0]), tensor([-1, -1, -1, 0, 0, 1]))
(tensor([9, 9, 0, 5, 4, 0]), tensor([-1, -1, -1, 4, 0, 1]))
(tensor([8, 9, 9, 6, 5, 8]), tensor([-1, -1, -1, 5, 8, 1]))
(tensor([4, 0, 8, 9, 9, 2]), tensor([-1, -1, -1, 9, 2, 1]))
...
这一下子可能看不清,我把第一条和第三条的 AutoRegress 过程示意图放在下面
99 + 32 = 131
------------------------
-1 -1 -1 1 3 1
9 9 3 2 1 3 1
55 + 3 = 58
------------------------
-1 -1 -1 8 5 0
5 5 0 3 8 5 0
注意设为 -1 的部分被 mask 掉不计算 loss(联系 3.2.2 节的 GPT forward 函数中损失计算部分)
class Trainer: @staticmethod def get_default_config(): C = CN() # device to train on C.device = 'auto' # dataloder parameters C.num_workers = 4 # optimizer parameters C.max_iters = None C.batch_size = 64 C.learning_rate = 3e-4 C.betas = (0.9, 0.95) C.weight_decay = 0.1 # only applied on matmul weights C.grad_norm_clip = 1.0 return C def __init__(self, config, model, train_dataset): self.config = config self.model = model self.optimizer = None self.train_dataset = train_dataset self.callbacks = defaultdict(list) # determine the device we'll train on if config.device == 'auto': self.device = 'cuda' if torch.cuda.is_available() else 'cpu' else: self.device = config.device self.model = self.model.to(self.device) print("running on device", self.device) # variables that will be assigned to trainer class later for logging and etc self.iter_num = 0 self.iter_time = 0.0 self.iter_dt = 0.0 def add_callback(self, onevent: str, callback): self.callbacks[onevent].append(callback) def set_callback(self, onevent: str, callback): self.callbacks[onevent] = [callback] def trigger_callbacks(self, onevent: str): for callback in self.callbacks.get(onevent, []): callback(self) def run(self): model, config = self.model, self.config # setup the optimizer self.optimizer = model.configure_optimizers(config) # setup the dataloader train_loader = DataLoader( self.train_dataset, sampler=torch.utils.data.RandomSampler(self.train_dataset, replacement=True, num_samples=int(1e10)), shuffle=False, pin_memory=True, batch_size=config.batch_size, num_workers=config.num_workers, ) model.train() self.iter_num = 0 self.iter_time = time.time() data_iter = iter(train_loader) while True: # fetch the next batch (x, y) and re-init iterator if needed try: batch = next(data_iter) except StopIteration: data_iter = iter(train_loader) batch = next(data_iter) batch = [t.to(self.device) for t in batch] x, y = batch # forward the model logits, self.loss = model(x, y) # backprop and update the parameters model.zero_grad(set_to_none=True) self.loss.backward() torch.nn.utils.clip_grad_norm_(model.parameters(), config.grad_norm_clip) self.optimizer.step() self.trigger_callbacks('on_batch_end') self.iter_num += 1 tnow = time.time() self.iter_dt = tnow - self.iter_time self.iter_time = tnow # termination conditions if config.max_iters is not None and self.iter_num >= config.max_iters: break
config.max_iters
,不断轮询训练集直到训练 batch 数量达标为止self.trigger_callbacks('on_batch_end')
,每个 batch 训练后调用它,这个函数的函数体将在训练主函数中实现minGPT/utils.py
文件中,请自行查阅def get_config(): C = CN() # system C.system = CN() C.system.seed = 3407 C.system.work_dir = './out/adder' # data C.data = AdditionDataset.get_default_config() # model C.model = GPT.get_default_config() C.model.model_type = 'gpt-nano' # trainer C.trainer = Trainer.get_default_config() C.trainer.learning_rate = 5e-4 # the model we're using is so small that we can go a bit faster return C
if __name__ == '__main__': # get default config and overrides from the command line, if any config = get_config() config.merge_from_args(sys.argv[1:]) print(config) setup_logging(config) set_seed(config.system.seed) # construct train and test datasets train_dataset = AdditionDataset(config.data, split='train') test_dataset = AdditionDataset(config.data, split='test') # construct the model config.model.vocab_size = train_dataset.get_vocab_size() config.model.block_size = train_dataset.get_block_size() model = GPT(config.model) # construct the trainer object config.trainer.max_iters = 5001 trainer = Trainer(config.trainer, model, train_dataset) train_losses = [] accuracies = [] # helper function for the evaluation of a model def eval_split(trainer, split, max_batches=None): dataset = {'train':train_dataset, 'test':test_dataset}[split] ndigit = config.data.ndigit results = [] mistakes_printed_already = 0 factors = torch.tensor([[10**i for i in range(ndigit+1)][::-1]]).to(trainer.device) loader = DataLoader(dataset, batch_size=100, num_workers=0, drop_last=False) for b, (x, y) in enumerate(loader): x = x.to(trainer.device) # isolate the first two digits of the input sequence alone d1d2 = x[:, :ndigit*2] # let the model sample the rest of the sequence d1d2d3 = model.generate(d1d2, ndigit+1, do_sample=False) # using greedy argmax, not sampling # isolate the last digit of the sampled sequence d3 = d1d2d3[:, -(ndigit+1):] d3 = d3.flip(1) # reverse the digits to their "normal" order # decode the integers from individual digits d1i = (d1d2[:,:ndigit] * factors[:,1:]).sum(1) d2i = (d1d2[:,ndigit:ndigit*2] * factors[:,1:]).sum(1) d3i_pred = (d3 * factors).sum(1) d3i_gt = d1i + d2i # manually calculate the ground truth # evaluate the correctness of the results in this batch correct = (d3i_pred == d3i_gt).cpu() # Software 1.0 vs. Software 2.0 fight RIGHT on this line haha for i in range(x.size(0)): results.append(int(correct[i])) if not correct[i] and mistakes_printed_already < 5: # only print up to 5 mistakes to get a sense mistakes_printed_already += 1 print("GPT claims that %d + %d = %d but gt is %d" % (d1i[i], d2i[i], d3i_pred[i], d3i_gt[i])) if max_batches is not None and b+1 >= max_batches: break rt = torch.tensor(results, dtype=torch.float) print("%s final score: %d/%d = %.2f%% correct" % (split, rt.sum(), len(results), 100*rt.mean())) if split == 'test': accuracies.append(100*rt.mean()) return rt.sum() # iteration callback top_score = 0 def batch_end_callback(trainer): global top_score train_losses.append(trainer.loss.item()) if trainer.iter_num % 10 == 0: print(f"iter_dt {trainer.iter_dt * 1000:.2f}ms; iter {trainer.iter_num}: train loss {trainer.loss.item():.5f}") if trainer.iter_num % 500 == 0: # evaluate both the train and test score train_max_batches = {1: None, 2: None, 3: 5}[config.data.ndigit] # if ndigit=2 we can afford the whole train set, ow no model.eval() with torch.no_grad(): train_score = eval_split(trainer, 'train', max_batches=train_max_batches) test_score = eval_split(trainer, 'test', max_batches=None) score = train_score + test_score # save the model if this is the best score we've seen so far if score > top_score: top_score = score print(f"saving model with new top score of {score}") ckpt_path = os.path.join(config.system.work_dir, "model.pt") torch.save(model.state_dict(), ckpt_path) # revert model to training mode model.train() trainer.set_callback('on_batch_end', batch_end_callback) # run the optimization trainer.run() # show performance fig, (ax1, ax2) = plt.subplots(2, 1, figsize=(8, 6)) x = np.arange(len(train_losses)) ax1.plot(x, train_losses, label='Train Loss') ax2.set_xlabel('iter') ax1.set_ylabel('Loss') ax1.legend() x = np.arange(len(accuracies)) * 500 ax2.plot(x, accuracies, label='Accuracy') ax2.set_xlabel('iter') ax2.set_ylabel('Accuracy') ax2.legend() plt.subplots_adjust(hspace=0.4) plt.show()
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。