A summarization website that can generate summaries from either YouTube videos or PDF files.
-
Updated
May 29, 2024 - Python
A summarization website that can generate summaries from either YouTube videos or PDF files.
Focus - Understanding contextual retrievability.
Fine-tuned Longformer for Summarization of Machine Learning Articles
This project was developed for a Kaggle competition focused on detecting Personally Identifiable Information (PII) in student writing. The primary objective was to build a robust model capable of identifying PII with high recall. The DeBERTa v3 transformer model was chosen for this task after comparing its performance with other transformer models.
Factuality check of the SemRep Predications
Longformer Encoder Decoder model for the legal domain, trained for long document abstractive summarization task.
An attempt of creating a model and pipeline for retrieving italian legal documents given a prompt from the user.
Kaggle NLP competition - Top 2% solution (36/2060)
This project applies the Longformer model to sentiment analysis using the IMDB movie review dataset. The Longformer model, introduced in "Longformer: The Long-Document Transformer," tackles long document processing with sliding-window and global attention mechanisms. The implementation leverages PyTorch, following the paper's architecture
Training and inference code for the claim veracity checker built on Longformer-4096 tuned to PUBHEALTH
This GitHub repository implements a novel approach for detecting Initial Public Offering (IPO) underpricing using pre-trained Transformers. The models, extended to handle large S-1 filings, leverage both textual information and financial indicators, outperforming traditional machine learning methods.
Industrial Text Scoring using Multimodal Deep Natural Language Processing 🚀 | Code for IEA AIE 2022 paper
A hyperpartisan news article classification system using BERT-based techniques. The goal was to leverage state-of-the-art transformer models like BERT, ROBERTa, and Longformer to accurately classify news articles as hyperpartisan or non-hyperpartisan.
Project as part of COMP34812: Natural Language Understanding
using transformers to do text classification.
A WebApp to summarize research papers using HuggingFace Transformers.
Convert pretrained RoBerta models to various long-document transformer models
[제 13회 투빅스 컨퍼런스] YoYAK - Yes or Yes, Attention with gap-sentence for Korean long sequence
Add a description, image, and links to the longformer topic page so that developers can more easily learn about it.
To associate your repository with the longformer topic, visit your repo's landing page and select "manage topics."