This repository is used to collect papers and code in the field of AI.
-
Updated
Jul 16, 2024
This repository is used to collect papers and code in the field of AI.
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Implementation of Vision Transformers (ViT) with a token merging mechanism
LinkOrgs: An R package for linking linking records on organizations using half a billion open-collaborated records from LinkedIn
This is the project repo associated with the paper "Disentangling and Integrating Relational and Sensory Information in Transformer Architectures" by Awni Altabaa, John Lafferty
This repository hosts BonyadAI, a Persian question answering AI Model. We developed an initial web crawler and scraper to gather the dataset. The second phase involved building a machine learning model based on word embeddings and NLP techniques. This AI model operates end-to-end, receiving user voice input and providing responses in Persian voice.
This repository contains the code for the Transformer-Representation Neural Topic Model (TNTM) based on the paper "Probabilistic Topic Modelling with Transformer Representations" by Arik Reuter, Anton Thielmann, Christoph Weisser, Benjamin Säfken and Thomas Kneib
An introduction to attention mechanisms and the vision transformer
Multi-Step Retrosynthesis Tool based on Augmented Disconnection Aware Triple Transformer Loop Predictions
An implementation of the base GPT-3 Model architecture from the paper by OPENAI "Language Models are Few-Shot Learners"
This project involves analyzing the relationship between Nvidia and AI technologies. The notebook covers data pre-processing steps, including importing necessary libraries and loading data. Further analysis and insights into Nvidia's impact on AI and related trends are provided.
toyGPT - Hands-on project for learning Transformer and GPT model
Basic Gesture Recognition Using mmWave Sensor - TI AWR1642
A numpy implementation of the Transformer model in "Attention is All You Need"
Official PyTorch implementation of the Vectorized Conditional Neural Field.
A simplistic pytorch implementation of LongVit using my previous implementation of LongNet as a foundation.
Welcome to quote our published papers, and the codes have been uploaded.
Yet Another Transformer Implementation
Educational code for understanding attention mechanisms. You will build a good intuition to K, Q, and V, key in modern Transformer architectures.
Add a description, image, and links to the transformer-architecture topic page so that developers can more easily learn about it.
To associate your repository with the transformer-architecture topic, visit your repo's landing page and select "manage topics."