训练期间 Huggingface 错误:AttributeError: 'str' object has no attribute 'size'

如何解决训练期间 Huggingface 错误:AttributeError: 'str' object has no attribute 'size'

在尝试使用 Pytorch Lightning 微调 Huggingface GPT2LMHeadModel 模型以进行休闲语言建模(给定单词序列,预测下一个单词)时,我在训练过程中遇到错误:

AttributeError: 'str' 对象没有属性 'size'

我们的训练代码出了什么问题?这是由于在 Pytorch DataCollatorForLanguageModeling 中错误使用了 DataLoader 吗?

可重现的示例:

import os
from pathlib import Path
import torch
import pytorch_lightning as pl
from transformers import (
    GPT2Config,GPT2LMHeadModel,GPT2Tokenizer,DataCollatorForLanguageModeling,)
from transformers.optimization import AdamW
from tokenizers import ByteLevelBPETokenizer
from torch.utils.data import (
    DataLoader,Dataset,)

TOKENIZER_DIRPATH = os.path.join("..","data")


def tokenize_data():
    tokenizer = ByteLevelBPETokenizer()
    tokenizer.train(
        files=os.path.join(TOKENIZER_DIRPATH,"words.txt"),vocab_size=50000,min_frequency=2,special_tokens=["<s>","</s>","<unk>","<mask>","<pad>",],)
    tokenizer.save_model("../data")


class MyDataset(Dataset):
    def __init__(self):
        tokenizer = GPT2Tokenizer(
            os.path.join(TOKENIZER_DIRPATH,"vocab.json"),os.path.join(TOKENIZER_DIRPATH,"merges.txt"),)

        src_file = Path(os.path.join(TOKENIZER_DIRPATH,"words.txt"))
        lines = src_file.read_text(encoding="utf-8").splitlines()
        self.examples = [tokenizer.encode(line) for line in lines]

    def __len__(self):
        return len(self.examples)

    def __getitem__(self,i):
        return torch.tensor(self.examples[i])


class MyDataModule(pl.LightningDataModule):
    def __init__(self):
        super().__init__()
        self.tokenizer = GPT2Tokenizer(
            os.path.join(TOKENIZER_DIRPATH,)

    def setup(self,stage):
        self.train_dataset = MyDataset()

    def train_dataloader(self):
        data_collator = DataCollatorForLanguageModeling(
            tokenizer=self.tokenizer,mlm=False
        )
        train_dataloader = DataLoader(self.train_dataset,collate_fn=data_collator)
        return train_dataloader


class MyModel(pl.LightningModule):
    def __init__(self,learning_rate,adam_beta1,adam_beta2,adam_epsilon):
        super().__init__()
        self.save_hyperparameters()
        config = GPT2Config()
        self.model = GPT2LMHeadModel(config)

    def forward(self,x):
        return self.model(x).logits

    def training_step(self,batch,batch_idx):
        input_ids,labels = batch
        loss = self.model(input_ids,labels=labels).loss
        self.log("train_loss",loss,on_epoch=True)
        return loss

    def configure_optimizers(self):
        optimizer = AdamW(
            self.parameters(),self.hparams.learning_rate,betas=(self.hparams.adam_beta1,self.hparams.adam_beta2),eps=self.hparams.adam_epsilon,)
        return optimizer


tokenize_data()
dm = MyDataModule()
model = MyModel(
    learning_rate=5e-5,adam_beta1=0.9,adam_beta2=0.999,adam_epsilon=1e-8,)

trainer = pl.Trainer()
trainer.fit(model,dm)

错误追溯:

Epoch 0:   0%|                                                                                                                                                                                                                                             | 0/9 [00:00<?,?it/s]
Traceback (most recent call last):
  File "test_gpt.py",line 102,in <module>
    trainer.fit(model,dm)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py",line 499,in fit
    self.dispatch()
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py",line 546,in dispatch
    self.accelerator.start_training(self)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/accelerators/accelerator.py",line 73,in start_training
    self.training_type_plugin.start_training(trainer)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/plugins/training_type/training_type_plugin.py",line 114,in start_training
    self._results = trainer.run_train()
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py",line 637,in run_train
    self.train_loop.run_training_epoch()
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 493,in run_training_epoch
    batch_output = self.run_training_batch(batch,batch_idx,dataloader_idx)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 655,in run_training_batch
    self.optimizer_step(optimizer,opt_idx,train_step_and_backward_closure)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 426,in optimizer_step
    model_ref.optimizer_step(
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/core/lightning.py",line 1387,in optimizer_step
    optimizer.step(closure=optimizer_closure)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/core/optimizer.py",line 214,in step
    self.__optimizer_step(*args,closure=closure,profiler_name=profiler_name,**kwargs)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/core/optimizer.py",line 134,in __optimizer_step
    trainer.accelerator.optimizer_step(optimizer,self._optimizer_idx,lambda_closure=closure,**kwargs)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/accelerators/accelerator.py",line 277,in optimizer_step
    self.run_optimizer_step(optimizer,lambda_closure,line 282,in run_optimizer_step
    self.training_type_plugin.optimizer_step(optimizer,lambda_closure=lambda_closure,**kwargs)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/plugins/training_type/training_type_plugin.py",line 163,in optimizer_step
    optimizer.step(closure=lambda_closure,**kwargs)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/transformers/optimization.py",line 318,in step
    loss = closure()
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 649,in train_step_and_backward_closure
    result = self.training_step_and_backward(
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 743,in training_step_and_backward
    result = self.training_step(split_batch,hiddens)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/trainer/training_loop.py",line 293,in training_step
    training_step_output = self.trainer.accelerator.training_step(args)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/accelerators/accelerator.py",line 156,in training_step
    return self.training_type_plugin.training_step(*args)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/pytorch_lightning/plugins/training_type/training_type_plugin.py",line 125,in training_step
    return self.lightning_module.training_step(*args,**kwargs)
  File "test_gpt.py",line 81,in training_step
    loss = self.model(input_ids,labels=labels).loss
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/torch/nn/modules/module.py",line 727,in _call_impl
    result = self.forward(*input,**kwargs)
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/transformers/models/gpt2/modeling_gpt2.py",line 904,in forward
    transformer_outputs = self.transformer(
  File "/opt/anaconda3/envs/test_huggingface/lib/python3.8/site-packages/torch/nn/modules/module.py",line 633,in forward
    input_shape = input_ids.size()
AttributeError: 'str' object has no attribute 'size'

Conda 包:

pytorch                   1.7.0           py3.8_cuda10.2.89_cudnn7.6.5_0    pytorch
pytorch-lightning         1.2.5              pyhd8ed1ab_0    conda-forge
tokenizers                0.10.1                   pypi_0    pypi
transformers              4.4.2                    pypi_0    pypi

解决方法

这里transformer最新版本可能会出现这个

pip install transformers==2.11.0

就我而言,这是可行的!!然后重启你的内核

版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。

相关推荐


使用本地python环境可以成功执行 import pandas as pd import matplotlib.pyplot as plt # 设置字体 plt.rcParams[&#39;font.sans-serif&#39;] = [&#39;SimHei&#39;] # 能正确显示负号 p
错误1:Request method ‘DELETE‘ not supported 错误还原:controller层有一个接口,访问该接口时报错:Request method ‘DELETE‘ not supported 错误原因:没有接收到前端传入的参数,修改为如下 参考 错误2:cannot r
错误1:启动docker镜像时报错:Error response from daemon: driver failed programming external connectivity on endpoint quirky_allen 解决方法:重启docker -&gt; systemctl r
错误1:private field ‘xxx‘ is never assigned 按Altʾnter快捷键,选择第2项 参考:https://blog.csdn.net/shi_hong_fei_hei/article/details/88814070 错误2:启动时报错,不能找到主启动类 #
报错如下,通过源不能下载,最后警告pip需升级版本 Requirement already satisfied: pip in c:\users\ychen\appdata\local\programs\python\python310\lib\site-packages (22.0.4) Coll
错误1:maven打包报错 错误还原:使用maven打包项目时报错如下 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-resources-plugin:3.2.0:resources (default-resources)
错误1:服务调用时报错 服务消费者模块assess通过openFeign调用服务提供者模块hires 如下为服务提供者模块hires的控制层接口 @RestController @RequestMapping(&quot;/hires&quot;) public class FeignControl
错误1:运行项目后报如下错误 解决方案 报错2:Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.1:compile (default-compile) on project sb 解决方案:在pom.
参考 错误原因 过滤器或拦截器在生效时,redisTemplate还没有注入 解决方案:在注入容器时就生效 @Component //项目运行时就注入Spring容器 public class RedisBean { @Resource private RedisTemplate&lt;String
使用vite构建项目报错 C:\Users\ychen\work&gt;npm init @vitejs/app @vitejs/create-app is deprecated, use npm init vite instead C:\Users\ychen\AppData\Local\npm-
参考1 参考2 解决方案 # 点击安装源 协议选择 http:// 路径填写 mirrors.aliyun.com/centos/8.3.2011/BaseOS/x86_64/os URL类型 软件库URL 其他路径 # 版本 7 mirrors.aliyun.com/centos/7/os/x86
报错1 [root@slave1 data_mocker]# kafka-console-consumer.sh --bootstrap-server slave1:9092 --topic topic_db [2023-12-19 18:31:12,770] WARN [Consumer clie
错误1 # 重写数据 hive (edu)&gt; insert overwrite table dwd_trade_cart_add_inc &gt; select data.id, &gt; data.user_id, &gt; data.course_id, &gt; date_format(
错误1 hive (edu)&gt; insert into huanhuan values(1,&#39;haoge&#39;); Query ID = root_20240110071417_fe1517ad-3607-41f4-bdcf-d00b98ac443e Total jobs = 1
报错1:执行到如下就不执行了,没有显示Successfully registered new MBean. [root@slave1 bin]# /usr/local/software/flume-1.9.0/bin/flume-ng agent -n a1 -c /usr/local/softwa
虚拟及没有启动任何服务器查看jps会显示jps,如果没有显示任何东西 [root@slave2 ~]# jps 9647 Jps 解决方案 # 进入/tmp查看 [root@slave1 dfs]# cd /tmp [root@slave1 tmp]# ll 总用量 48 drwxr-xr-x. 2
报错1 hive&gt; show databases; OK Failed with exception java.io.IOException:java.lang.RuntimeException: Error in configuring object Time taken: 0.474 se
报错1 [root@localhost ~]# vim -bash: vim: 未找到命令 安装vim yum -y install vim* # 查看是否安装成功 [root@hadoop01 hadoop]# rpm -qa |grep vim vim-X11-7.4.629-8.el7_9.x
修改hadoop配置 vi /usr/local/software/hadoop-2.9.2/etc/hadoop/yarn-site.xml # 添加如下 &lt;configuration&gt; &lt;property&gt; &lt;name&gt;yarn.nodemanager.res