COLLECTED BY
Organization:
Internet Archive
Focused crawls are collections of frequently-updated webcrawl data from narrow (as opposed to broad or wide) web crawls, often focused on a single domain or subdomain.
The Wayback Machine - https://web.archive.org/web/20200626181653/https://github.com/topics/self-attention
Here are
85 public repositories
matching this topic...
Updated
Apr 12, 2020
Python
Datasets, tools, and benchmarks for representation learning of code.
Updated
Jun 22, 2020
Jupyter Notebook
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
Updated
Jan 1, 2019
Python
CCNet: Criss-Cross Attention for Semantic Segmentation (ICCV 2019).
Updated
Jan 30, 2020
Python
Text classification using deep learning models in Pytorch
Updated
Nov 17, 2018
Python
A PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.
Updated
May 7, 2020
Python
A Structured Self-attentive Sentence Embedding
Updated
Sep 22, 2019
Python
Keras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Updated
Mar 11, 2020
Jupyter Notebook
An implementation of DeepMind's Relational Recurrent Neural Networks in PyTorch.
Updated
Dec 27, 2018
Python
Implementing Stand-Alone Self-Attention in Vision Models using Pytorch
Updated
Feb 13, 2020
Python
Tensorflow implementation of "A Structured Self-Attentive Sentence Embedding"
Updated
Jan 28, 2020
Python
Neat (Neural Attention) Vision, is a visualization tool for the attention mechanisms of deep-learning models for Natural Language Processing (NLP) tasks. (framework-agnostic)
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
Updated
Dec 10, 2018
Python
Updated
Feb 26, 2019
Python
A variant of the Self Attention GAN named: FAGAN (Full Attention GAN)
Updated
Nov 21, 2018
Python
Transformer for Graph Classification (in Pytorch and Tensorflow)
Updated
Jun 24, 2020
Python
Simple Tensorflow Implementation of "A Structured Self-attentive Sentence Embedding" (ICLR 2017)
Updated
Jun 25, 2018
Python
Updated
Nov 3, 2018
Python
Multi-turn dialogue baselines written in PyTorch
Updated
Mar 10, 2020
Python
The implementation of DeBERTa
Updated
Jun 26, 2020
Python
Representation learning on dynamic graphs using self-attention networks
Updated
Feb 11, 2020
Python
Updated
Apr 15, 2019
Python
Updated
May 5, 2019
Python
Code for Describing a Knowledge Base
Updated
Feb 12, 2020
Python
Sequence to Sequence Models in PyTorch
Updated
Jun 26, 2020
Python
Transformer-based Relational Memory for Knowledge Graph Embeddings (ACL 2020) (in Pytorch and Tensorflow)
Updated
Jun 22, 2020
Python
PyTorch implementation of the model presented in "Satellite Image Time Series Classification with Pixel-Set Encoders and Temporal Self-Attention"
Updated
Mar 30, 2020
Python
Chatbot using Tensorflow (Model is transformer) ko
Updated
Dec 10, 2018
Python
Collection of important Question-Answering papers implemented in PyTorch
Updated
Jun 25, 2020
Jupyter Notebook
Improve this page
Add a description, image, and links to the
self-attention
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
self-attention
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.
In official documents, there is a notice 'Each function object is meant to be used only once (in the forward pass).' in subclass of torch.autograd.Function.
In model SpGraphAttentionLayer, you have use the object of SpecialSpmmFunction(self.special_spmm) twice, one for e_rowsum and one for h_prime.
Is it the right usage for subclass of torch.autograd.Function?