Orchestrate Swarms of Agents From Any Framework Like OpenAI, Langchain, and Etc for Business Operation Automation. Join our Community: https://discord.gg/DbjBMJTSWD
-
Updated
Jun 13, 2024 - Python
Orchestrate Swarms of Agents From Any Framework Like OpenAI, Langchain, and Etc for Business Operation Automation. Join our Community: https://discord.gg/DbjBMJTSWD
An attention based approach to convert Indian Sign Language to Text using simulated hand gesture data
Source code for the GAtt method in "Revisiting Attention Weights as Interpretations of Message-Passing Neural Networks".
This project uses PyTorch to classify bone fractures. As well as fine-tuning some famous CNN architectures (like VGG 19, MobileNetV3, RegNet,...), we designed our own architecture. Additionally, we used Transformer architectures (such as Vision Transformer and Swin Transformer). This dataset is Bone Fracture Multi-Region X-ray, available on Kaggle.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Dataset and scripts for paper "A Human-Annotated Video Dataset for Training and Evaluation of 360-Degree Video Summarization Methods", Proc. 1st Int. Workshop on Video for Immersive Experiences (Video4IMX-2024) at ACM IMX 2024, Stockholm, Sweden, June 2024.
[CVPR 2024] "CFAT: Unleashing Triangular Windows for Image Super-resolution"
Experimental project on building custom LSTM and LSTM with Attention layer for comparison analysis on FTS forecasting (June 2024)
A minimal implementation of a denoising diffusion model in PyTorch.
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Researching causal relationships in time series data using Temporal Convolutional Networks (TCNs) combined with attention mechanisms. This approach aims to identify complex temporal interactions. Additionally, we're incorporating uncertainty quantification to enhance the reliability of our causal predictions.
An extension for the code and data of the paper "Human Choice Prediction in Language-based Non-Cooperative Games: Simulation-based Off-Policy Evaluation" (Shapira et al. 2023). This project was conducted by Yogev Namir and Avishag Nevo.
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
A simple but complete full-attention transformer with a set of promising experimental features from various papers
Xllama🦙 is an Extensible advanced language model framework, inspired by the original Llama model.
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
QuillGPT is an implementation of the GPT decoder block based on the architecture from Attention is All You Need paper by Vaswani et. al. in PyTorch. Additionally, this repository contains two pre-trained models — Shakespearean GPT and Harpoon GPT, a Streamlit Playground, Containerized FastAPI Microservice, training - inference scripts & notebooks.
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
Add a description, image, and links to the attention-mechanism topic page so that developers can more easily learn about it.
To associate your repository with the attention-mechanism topic, visit your repo's landing page and select "manage topics."