Simple State-of-the-Art BERT-Based Sentence Classification with Keras / TensorFlow 2. Built with HuggingFace's Transformers.
-
Updated
May 26, 2024 - Python
Simple State-of-the-Art BERT-Based Sentence Classification with Keras / TensorFlow 2. Built with HuggingFace's Transformers.
Triple Branch BERT Siamese Network for fake news classification on LIAR-PLUS dataset in PyTorch
TunBERT is the first release of a pre-trained BERT model for the Tunisian dialect using a Tunisian Common-Crawl-based dataset. TunBERT was applied to three NLP downstream tasks: Sentiment Analysis (SA), Tunisian Dialect Identification (TDI) and Reading Comprehension Question-Answering (RCQA)
BERT which stands for Bidirectional Encoder Representations from Transformations is the SOTA in Transfer Learning in NLP.
This is the code for loading the SenseBERT model, described in our paper from ACL 2020.
The code of Team Rhinobird for Mining the Web of HTML-embedded Product Data Task One at ISWC2020
BERT implementation for radiology full-text reports
Part-of-Speech Tagging for simplified and traditional Chinese data with BERT & RoBERTa
Quick and easy tutorial to serve HuggingFace sentiment analysis model using torchserve
Code and data for the NLLP 2021 paper: `Multi-granular Legal topic Classification on Greek Legislation`
[PyPI] BERT Word Embeddings
Comparing between residual stream and highway stream in transformers(BERT) .
B.Sc. Thesis Deep Learning & NLP research on Medical Image Captioning
Fine-tuning framework for BERT like models on RACE
This repository contains NLP Transfer learning projects with deployment and integration with UI.
A dark web analysis tool.
Code and models for the paper 'Exploring Multi-Modal Representations for Ambiguity Detection & Coreference Resolution in the SIMMC 2.0 Challenge' published at AAAI 2022 DSTC10 Workshop
BERTs based rank relative ratings of toxicity between comments
A Chinese idiom recommendation system based on BERT pre-training language model.
Add a description, image, and links to the bert-models topic page so that developers can more easily learn about it.
To associate your repository with the bert-models topic, visit your repo's landing page and select "manage topics."