Paper Reading: Modeling Relational Data with Graph Convolutional Networks

This paper introduce Relational Graph Convolutional Networks (R-GCNs) which deal with relational data. In addition, techniques for parameter sharing and to enforce sparsity constraints are introduced to apply R-GCNs to multigraphs with large numbers of relations. 1. Relational Graph Convolutional Networks Given directed and labeled multi-graphs as with nodes and labeled edges , where is a relation type. In a relational multi-graph, R-GCNs calculates the … Continue reading Paper Reading: Modeling Relational Data with Graph Convolutional Networks

Paper Reading: What Does BERT Look At? An Analysis of BERT’s Attention

code of this paper: link High-level Summary This paper studies the attention maps of the pre-trained BERT-base model. More specically, it : explore generally how BERT’s attention heads behave. eg. attending to fixed positional offsets. eg. attending broadly over the whole sentence. a large amount of attention attends to [SEP]. attention heads in the same layer behave similarly. probe each attention head for linguistic phenomena. … Continue reading Paper Reading: What Does BERT Look At? An Analysis of BERT’s Attention

Paper Reading: Universal Text Representation from BERT: An Empirical Study

This paper studies the layer-wise BERT activations for sentence-level tasks and passage-level tasks. 1. BERT Sentence Embedding SentEval toolkit is used to evaluate the quality of sentence representations from BERT activations. It has a variety of downstreaming sentence-level tasks and probing tasks. More details about SentEval are at: https://github.com/facebookresearch/SentEval 1.1 [CLS] from different layers [CLS] token embeddings from different layers are used for classification. Only … Continue reading Paper Reading: Universal Text Representation from BERT: An Empirical Study

Paper Reading: Revealing the Dark Secrets of BERT

The paper tries to answer the following questions: What are the common attention patterns, how do they change during fine-tuning, and how does that impact the performance on a given task?  What linguistic knowledge is encoded in self-attention weights of the fine-tuned models and what portion of it comes from the pre-trained BERT? How different are the self-attention patterns of different heads, and how important … Continue reading Paper Reading: Revealing the Dark Secrets of BERT

Paper Reading: CoSimLex: A Resource for Evaluating Graded Word Similarity in Context

Introduction Previous methods to evaluate word embeddings intrinsiclly (e.g. WordSim-353, SimLex-999, word analogy task) ignore the context and treat words in isolation. This paper proposes a dataset CoSimLex to evaluate the ability of word embeddings that reflect similarity judgements in context and answer the following question: How well do word embeddings model the effects that context has on word meaning? CoSimLex is used as the … Continue reading Paper Reading: CoSimLex: A Resource for Evaluating Graded Word Similarity in Context

Paper Reading: Do Attention Heads in BERT Track Syntactic Dependencies?

The paper specifically studies the ability of attention heads(of BERT-like models) that can recover syntactic dependency relations. Method 1: Maximum Attention Weights (MAX) For a given token A, a token B that has the highest attention weight with respect to the token A should be related to token A. A relation is assigned to such that   for each row i where is the attention weights … Continue reading Paper Reading: Do Attention Heads in BERT Track Syntactic Dependencies?

Free tools for collecting data from social network

BackTweets BackTweets is a twitter time machine which enables you to search through a tweet history for tweets that link back to your site. Website: http://backtweets.com/ Icerocket Enter keywords to see mentions trended over time. Website: http://www.icerocket.com/ Webhose Website: https://webhose.io Python library: https://github.com/Buzzilla/webhose-python Social Searcher Export to csv files easily. Website: https://www.social-searcher.com/ TipTop Website: http://feeltiptop.com/ search results are grouped into images, videos, and sentiment polarities. Twazzup real-time content from Twitter Website: http://twazzup.com/ … Continue reading Free tools for collecting data from social network

Common Knowledge Base

Recently, I read a lot of papers making use of Wikipedia and also other different knowledge bases. Therefore, I collect some common ones which may be helpful for future use. ConceptNet ConceptNet is a semantic network containing lots of things computers should know about the world, especially when understanding text written by people. WordNet a lexical knowledge base of about 25,000 words grouped into an … Continue reading Common Knowledge Base