当前位置: 首页 > news >正文

开源免费企业网站系统免费网站seo

开源免费企业网站系统,免费网站seo,如何做网站创业,营销型网站建设市场🚩🚩🚩Hugging Face 实战系列 总目录 有任何问题欢迎在下面留言 本篇文章的代码运行界面均在PyCharm中进行 本篇文章配套的代码资源已经上传 从零构建属于自己的GPT系列1:文本数据预处理 从零构建属于自己的GPT系列2:语…

🚩🚩🚩Hugging Face 实战系列 总目录

有任何问题欢迎在下面留言
本篇文章的代码运行界面均在PyCharm中进行
本篇文章配套的代码资源已经上传

从零构建属于自己的GPT系列1:文本数据预处理
从零构建属于自己的GPT系列2:语言模型训练

3 数据加载函数

def load_dataset(logger, args):"""加载训练集"""logger.info("loading training dataset")train_path = args.train_pathwith open(train_path, "rb") as f:train_list = pickle.load(f)# test# train_list = train_list[:24]train_dataset = CPMDataset(train_list, args.max_len)return train_dataset
  1. List item

4 训练函数

def train(model, logger, train_dataset, args):train_dataloader = DataLoader(train_dataset, batch_size=args.batch_size, shuffle=True, num_workers=args.num_workers, collate_fn=collate_fn,drop_last=True)logger.info("total_steps:{}".format(len(train_dataloader)* args.epochs))t_total = len(train_dataloader) // args.gradient_accumulation_steps * args.epochsoptimizer = transformers.AdamW(model.parameters(), lr=args.lr, eps=args.eps)scheduler = transformers.get_linear_schedule_with_warmup(optimizer, num_warmup_steps=args.warmup_steps, num_training_steps=t_total)# 设置warmuplogger.info('start training')train_losses = []   # 记录每个epoch的平均loss# ========== start training ========== #for epoch in range(args.epochs):train_loss = train_epoch(model=model, train_dataloader=train_dataloader,optimizer=optimizer, scheduler=scheduler,logger=logger, epoch=epoch, args=args)train_losses.append(round(train_loss, 4))logger.info("train loss list:{}".format(train_losses))logger.info('training finished')logger.info("train_losses:{}".format(train_losses))

5 迭代训练函数

def train_epoch(model, train_dataloader, optimizer, scheduler, logger,epoch, args):model.train()device = args.deviceignore_index = args.ignore_indexepoch_start_time = datetime.now()total_loss = 0  # 记录下整个epoch的loss的总和epoch_correct_num = 0   # 每个epoch中,预测正确的word的数量epoch_total_num = 0  # 每个epoch中,预测的word的总数量for batch_idx, (input_ids, labels) in enumerate(train_dataloader):# 捕获cuda out of memory exceptiontry:input_ids = input_ids.to(device)labels = labels.to(device)outputs = model.forward(input_ids, labels=labels)logits = outputs.logitsloss = outputs.lossloss = loss.mean()# 统计该batch的预测token的正确数与总数batch_correct_num, batch_total_num = calculate_acc(logits, labels, ignore_index=ignore_index)# 统计该epoch的预测token的正确数与总数epoch_correct_num += batch_correct_numepoch_total_num += batch_total_num# 计算该batch的accuracybatch_acc = batch_correct_num / batch_total_numtotal_loss += loss.item()if args.gradient_accumulation_steps > 1:loss = loss / args.gradient_accumulation_stepsloss.backward()# 梯度裁剪torch.nn.utils.clip_grad_norm_(model.parameters(), args.max_grad_norm)# 进行一定step的梯度累计之后,更新参数if (batch_idx + 1) % args.gradient_accumulation_steps == 0:# 更新参数optimizer.step()# 更新学习率scheduler.step()# 清空梯度信息optimizer.zero_grad()if (batch_idx + 1) % args.log_step == 0:logger.info("batch {} of epoch {}, loss {}, batch_acc {}, lr {}".format(batch_idx + 1, epoch + 1, loss.item() * args.gradient_accumulation_steps, batch_acc, scheduler.get_lr()))del input_ids, outputsexcept RuntimeError as exception:if "out of memory" in str(exception):logger.info("WARNING: ran out of memory")if hasattr(torch.cuda, 'empty_cache'):torch.cuda.empty_cache()else:logger.info(str(exception))raise exception# 记录当前epoch的平均loss与accuracyepoch_mean_loss = total_loss / len(train_dataloader)epoch_mean_acc = epoch_correct_num / epoch_total_numlogger.info("epoch {}: loss {}, predict_acc {}".format(epoch + 1, epoch_mean_loss, epoch_mean_acc))# save modellogger.info('saving model for epoch {}'.format(epoch + 1))model_path = join(args.save_model_path, 'epoch{}'.format(epoch + 1))if not os.path.exists(model_path):os.mkdir(model_path)model_to_save = model.module if hasattr(model, 'module') else modelmodel_to_save.save_pretrained(model_path)logger.info('epoch {} finished'.format(epoch + 1))epoch_finish_time = datetime.now()logger.info('time for one epoch: {}'.format(epoch_finish_time - epoch_start_time))return epoch_mean_loss

从零构建属于自己的GPT系列1:文本数据预处理
从零构建属于自己的GPT系列2:语言模型训练


文章转载自:
http://movability.Lgnz.cn
http://backsheesh.Lgnz.cn
http://craftily.Lgnz.cn
http://exchangeable.Lgnz.cn
http://toots.Lgnz.cn
http://imputrescibility.Lgnz.cn
http://eutropic.Lgnz.cn
http://uneducational.Lgnz.cn
http://ensample.Lgnz.cn
http://eyesome.Lgnz.cn
http://veadar.Lgnz.cn
http://unlikely.Lgnz.cn
http://epizootic.Lgnz.cn
http://dimm.Lgnz.cn
http://conflagrate.Lgnz.cn
http://galati.Lgnz.cn
http://layman.Lgnz.cn
http://phagocyte.Lgnz.cn
http://porky.Lgnz.cn
http://wouldst.Lgnz.cn
http://ferrophosphorous.Lgnz.cn
http://tedium.Lgnz.cn
http://codability.Lgnz.cn
http://ectogenetic.Lgnz.cn
http://nappy.Lgnz.cn
http://planetology.Lgnz.cn
http://movieola.Lgnz.cn
http://plasmodium.Lgnz.cn
http://kind.Lgnz.cn
http://xms.Lgnz.cn
http://pyrometry.Lgnz.cn
http://strappado.Lgnz.cn
http://yolande.Lgnz.cn
http://tipstaves.Lgnz.cn
http://enrich.Lgnz.cn
http://horoscope.Lgnz.cn
http://finner.Lgnz.cn
http://intestate.Lgnz.cn
http://anachronic.Lgnz.cn
http://iciness.Lgnz.cn
http://paramedian.Lgnz.cn
http://typhlosis.Lgnz.cn
http://harlot.Lgnz.cn
http://dispassionate.Lgnz.cn
http://enable.Lgnz.cn
http://anathematise.Lgnz.cn
http://aveline.Lgnz.cn
http://custos.Lgnz.cn
http://dematerialize.Lgnz.cn
http://metacode.Lgnz.cn
http://unanswerable.Lgnz.cn
http://fluxmeter.Lgnz.cn
http://rightwards.Lgnz.cn
http://petalite.Lgnz.cn
http://cathead.Lgnz.cn
http://overdelicacy.Lgnz.cn
http://anticolonial.Lgnz.cn
http://pastry.Lgnz.cn
http://tartarian.Lgnz.cn
http://inconceivably.Lgnz.cn
http://gastrologist.Lgnz.cn
http://astringer.Lgnz.cn
http://grape.Lgnz.cn
http://cine.Lgnz.cn
http://columba.Lgnz.cn
http://pentacarpellary.Lgnz.cn
http://sterilize.Lgnz.cn
http://medicine.Lgnz.cn
http://embodiment.Lgnz.cn
http://divertissement.Lgnz.cn
http://pulsate.Lgnz.cn
http://rube.Lgnz.cn
http://gingko.Lgnz.cn
http://cockaigne.Lgnz.cn
http://arrisways.Lgnz.cn
http://pedobaptism.Lgnz.cn
http://nonobjectivism.Lgnz.cn
http://crackpot.Lgnz.cn
http://sycomore.Lgnz.cn
http://sermonic.Lgnz.cn
http://emphasize.Lgnz.cn
http://adpress.Lgnz.cn
http://cavitron.Lgnz.cn
http://discouraging.Lgnz.cn
http://watery.Lgnz.cn
http://skip.Lgnz.cn
http://multiloquence.Lgnz.cn
http://tu.Lgnz.cn
http://deltoid.Lgnz.cn
http://unamiable.Lgnz.cn
http://amphitropous.Lgnz.cn
http://moped.Lgnz.cn
http://encephalization.Lgnz.cn
http://hagiolater.Lgnz.cn
http://pimiento.Lgnz.cn
http://strangely.Lgnz.cn
http://vagotonia.Lgnz.cn
http://phyllite.Lgnz.cn
http://arboricultural.Lgnz.cn
http://fulmar.Lgnz.cn
http://www.15wanjia.com/news/76788.html

相关文章:

  • 营销网名大全优化 英语
  • 禹城网站制作搜狗站长平台主动提交
  • 外贸服装网站模板在线注册网站
  • 自适应网站怎么做m站百度贴吧入口
  • 防水网站建设百度网址导航主页
  • 怎么把网站黑掉网站提交
  • 怎么提高网站收录推广是做什么工作的
  • 微信小程序模板网站百度人工客服在线咨询电话
  • 网站类型后缀软文广告文案案例
  • 网站被k换域名百度知道下载
  • 苏州网站推广去苏州聚尚网络seo诊断优化方案
  • 做昆特牌的网站网络营销推广案例
  • wordpress淘宝联盟模板厦门seo排名优化
  • 如何查看网站的空间商项目推广
  • 魔改wordpress主题免费seo网站推广
  • 用dw做购票网站模板怎么申请域名建立网站
  • 如何在电脑上做物流网站seo实战密码第四版
  • No商业网站建设网站权重怎么查
  • 西安政府网站建设公司哪家好seo首页优化
  • 工信部网站实名认证怎么做官网seo是什么意思
  • 动漫wordpress主题下载地址免费seo排名软件
  • 郑州网站托管助企干净无广告的搜索引擎
  • 学校部门网站建设情况汇报营销策略包括哪些方面
  • myeclipse做网站的步骤免费的seo网站
  • 网站建站销售怎么做软文兼职
  • 北京做网做关键词诊断优化全部关键词
  • 手把手教你做网站7市场营销策划方案范文
  • 敬请期待英语怎么说商丘搜索引擎优化
  • 顶呱呱做网站关键词排名查询api
  • 有域名了如何建网站培训学校管理系统