Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
-
Updated
Nov 6, 2020 - Python
{{ message }}
Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS, 海量中文预训练ALBERT模型
RoBERTa中文预训练模型: RoBERTa for Chinese
中文语言理解基准测评 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
news-please - an integrated web crawler and information extractor for news that just works.
Integrating the Best of TF into PyTorch, for Machine Learning, Natural Language Processing, and Text Generation. This is part of the CASL project: http://casl-project.ai/
CLUENER2020 中文细粒度命名实体识别 Fine Grained Named Entity Recognition
高质量中文预训练模型集合:最先进大模型、最快小模型、相似度专门模型
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
PhoBERT: Pre-trained language models for Vietnamese (EMNLP-2020 Findings)
BERTweet: A pre-trained language model for English Tweets (EMNLP-2020)
中文预训练XLNet模型: Pre-Trained Chinese XLNet_Large
Official implementation of the paper “GECToR – Grammatical Error Correction: Tag, Not Rewrite” // Published on BEA15 Workshop (co-located with ACL 2020) https://www.aclweb.org/anthology/2020.bea-1.16.pdf
Simple State-of-the-Art BERT-Based Sentence Classification with Keras / TensorFlow 2. Built with HuggingFace's Transformers.
One-Stop Solution to encode sentence to fixed length vectors from various embedding techniques
transform multi-label classification as sentence pair task, with more training data and information
IMPORTANT: this repository statys for HISTORICAL reasons only. The actively supported repository is "openroberta-lab". In the "develop" branch of this repository please read the files "IMPORTANT_CHANGE.md" and "README.md". Thank you very much.
A Dutch RoBERTa-based language model
The implementation of DeBERTa
Tensorflow and Keras implementation of the state of the art researches in Dialog System NLU
All other language models do not perform as well as HappyROBERTA large for masked word prediction. We should encourage users to use HappyROBERTA Large by displaying a logger message if they use a suboptimal language model. This message will encourage them to use HappyROBERTA Large.
There are still some situations where a user may want to use another model, so we will keep them available.
BOND: BERT-Assisted Open-Domain Name Entity Recognition with Distant Supervision
reference pytorch code for named entity tagging
this is roberta wwm base distilled model which was distilled from roberta wwm by roberta wwm large
create an ardu program with 2 buttons allocated to the pin (inconsistency)
Models to perform neural summarization (extractive and abstractive) using machine learning transformers and a tool to convert abstractive summarization datasets to the extractive task.
Add a description, image, and links to the roberta topic page so that developers can more easily learn about it.
To associate your repository with the roberta topic, visit your repo's landing page and select "manage topics."
Hi, I am so interesting in your project, and wonder if you need contributor and how could I make my own contribution?