机器学习知识交换所(MLKX)

我在github上建了一个知识索引列表,以后会不断的刷新和结构化它。请点击此处访问我的列表。我的目的是不断的吸收各种机器学习,特别是深度学习的知识,构建称为机器学习的知识体系,然后我有另外一个Python的项目去构建一个文本阅读机器人(目前还在起步阶段)。

Deep Machine Learning Knowledge Exchange

Hello, this is winnerineast. I believe the better future is Human Being + Machine and I’m working on it in order to make it happen. Here is the inventory for all kinds of knowledges I collected from internet without any sign-in.

Special and equivalent thanks to (I just appended the name at the tail when I leverage or borrow his/her information)

Andrew Thomas

Keon Kim

Nam Vu

Denny Britz

Flood Sung

Part of the informations are from

ai-reading-list

nlp-reading-group

awesome-spanish-nlp

jjangsangy’s awesome-nlp

awesome-machine-learning

DL4NLP

Notes and Tutorials

Gernal Machine Learning Topics

Comprehensive list of data science resources

DigitalMind’s Artificial Intelligence resources

Awesome Machine Learning

CreativeAi’s Machine Learning

Machine Learning Blogby Brian McFee

Machine Learning in a Week

Machine Learning in a Year

How I wrote my first Machine Learning program in 3 days

Learning Path : Your mentor to become a machine learning expert

You Too Can Become a Machine Learning Rock Star! No PhD

How to become a Data Scientist in 6 months: A hacker’s approach to career planning

Video

Slide

5 Skills You Need to Become a Machine Learning Engineer

Are you a self-taught machine learning engineer? If yes, how did you do it & how long did it take you?

How can one become a good machine learning engineer?

A Learning Sabbatical focused on Machine Learning

Algorithms

10 Machine Learning Algorithms Explained to an ‘Army Soldier’

Top 10 data mining algorithms in plain English

10 Machine Learning Terms Explained in Simple English

A Tour of Machine Learning Algorithms

The 10 Algorithms Machine Learning Engineers Need to Know

Comparing supervised learning algorithms

Machine Learning Algorithms: A collection of minimal and clean implementations of machine learning algorithms

Interview Machine Learning Engineer Questions

How To Prepare For A Machine Learning Interview

40 Interview Questions asked at Startups in Machine Learning / Data Science

21 Must-Know Data Science Interview Questions and Answers

Top 50 Machine learning Interview questions & Answers

Machine Learning Engineer interview questions

Popular Machine Learning Interview Questions

What are some common Machine Learning interview questions?

What are the best interview questions to evaluate a machine learning researcher?

Collection of Machine Learning Interview Questions

121 Essential Machine Learning Questions & Answers

NLP

General

Deep Learning, NLP, and Representations

Word2Vec

Relation Extraction with Matrix Factorization and Universal Schemas

Towards a Formal Distributional Semantics: Simulating Logical Calculi with Tensors

Presentation slides for MLN tutorial

Presentation slides for QA applications of MLNs

Presentation slides

Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations

Blog Post onDeep Learning, NLP, and Representations

Blog Post onNLP Tutorial

Natural Language Processing Blogby Hal Daumé III

Word Vectors

Word2Vec tutorialinTensorFlow

Andrew Thomas notes on neural networks

Word2vec Parameter Learning Explained

The amazing power of word vectors

GloVe: Global vectors for word representation

Evalutaion section led to controversy

Glove source code and training data

Sentiment Analysis

doc2vec tutorial

FastText blog

seq2seq tutorialinTensorFlow.

Learning machine learning hints and clues

Can I learn and get a job in Machine Learning without studying CS Master and PhD?

How do I get a job in Machine Learning as a software programmer who self-studies Machine Learning, but never has a chance to use it at work?

What skills are needed for machine learning jobs?

There are two sides to machine learning:

I think the best way for practice-focused methodology is something like‘practice — learning — practice’, that means where students first come with some existing projects with problems and solutions (practice) to get familiar with traditional methods in the area and perhaps also with their methodology. After practicing with some elementary experiences, they can go into the books and study the underlying theory, which serves to guide their future advanced practice and will enhance their toolbox of solving practical problems. Studying theory also further improves their understanding on the elementary experiences, and will help them acquire advanced experiences more quickly.

Nam Vu – Top-down learning path: machine learning for software engineers|

What if I’m Not Good at Mathematics

5 Techniques To Understand Machine Learning Algorithms Without the Background in Mathematics

How do I learn machine learning?

What is the difference between Data Analytics, Data Analysis, Data Mining, Data Science, Machine Learning, and Big Data?

Learning How to Learn

Don’t Break The Chain

How to learn on your own

Kaggle knowledge competitions

Kaggle Competitions: How and where to begin?

How a Beginner Used Small Projects To Get Started in Machine Learning and Compete on Kaggle

Master Kaggle By Competing Consistently

Courses

GeneralMachine Learning Topics

A Visual Introduction to Machine Learning

A Gentle Guide to Machine Learning

Introduction to Machine Learning for Developers

Machine Learning basics for a newbie

How do you explain Machine Learning and Data Mining to non Computer Science people?

Machine Learning: Under the hood. Blog post explains the principles of machine learning in layman terms. Simple and clear

What is machine learning, and how does it work?

Deep Learning – A Non-Technical Introduction

The Machine Learning Mastery Method

Machine Learning for Programmers

Applied Machine Learning with Machine Learning Mastery

Python Machine Learning Mini-Course

Machine Learning Algorithms Mini-Course

Machine learning is fun

Machine Learning is Fun!

Part 2: Using Machine Learning to generate Super Mario Maker levels

Part 3: Deep Learning and Convolutional Neural Networks

Part 4: Modern Face Recognition with Deep Learning

Part 5: Language Translation with Deep Learning and the Magic of Sequences

Inky Machine Learning

Part 1: What is Machine Learning ?

Part 2: Supervised Learning and Unsupervised Learning

Machine learning: an in-depth, non-technical guide

Overview, goals, learning types, and algorithms

Data selection, preparation, and modeling

Model evaluation, validation, complexity, and improvement

Model performance and error analysis

Unsupervised learning, related fields, and machine learning in practice

Video Series

Machine Learning for Hackers

Fresh Machine Learning

Machine Learning Recipes with Josh Gordon

Everything You Need to know about Machine Learning in 30 Minutes or Less

A Friendly Introduction to Machine Learning

Nuts and Bolts of Applying Deep Learning – Andrew Ng

BigML Webinar

Video

Resources

mathematicalmonk’s Machine Learning tutorials

Machine learning in Python with scikit-learn

GitHub repository

Blog

My playlist – Top YouTube Videos on Machine Learning, Neural Network & Deep Learning

16 New Must Watch Tutorials, Courses on Machine Learning

DeepLearning.TV

Learning To See

MOOC

Udacity’s Intro to Machine Learning

Udacity Intro to Machine Learning Review

Udacity’s Supervised, Unsupervised & Reinforcement

Machine Learning Foundations: A Case Study Approach

Coursera’s Machine Learning

Video only

Coursera Machine Learning review

Coursera: Machine Learning Roadmap

Machine Learning Distilled

BigML training

Coursera’s Neural Networks for Machine Learning

Taught by Geoffrey Hinton, a pioneer in the field of neural networks

Machine Learning - CS - Oxford University

Creative Applications of Deep Learning with TensorFlow

Intro to Descriptive Statistics

Intro to Inferential Statistics

Resources

Learn Machine Learning in a Single Month

The Non-Technical Guide to Machine Learning & Artificial Intelligence

Machine Learning for Software Engineers on Hacker News

Machine Learning for Developers

Machine Learning Advice for Developers

Machine Learning For Complete Beginners

Getting Started with Machine Learning: For absolute beginners and fifth graders

How to Learn Machine Learning: The Self-Starter Way

Machine Learning Self-study Resources

Level-Up Your Machine Learning

Enough Machine Learning to Make Hacker News Readable Again

Video

Slide

Dive into Machine Learning

Machine Learning courses in Universities

Stanford

Machine Learning Summer Schools

Oxford

Cambridge

NLP

Kyunghyun Cho’s NLP course in NYU

Stanford Natural Language ProcessingIntro NLP course with videos. This hasno deep learning. But it is a good primer for traditional nlp.

Stanford CS 224D: Deep Learning for NLP class

Richard Socher. (2016) Class with syllabus, and slides. Videos:2015 lectures/2016 lectures

Michael Collins– one of the best NLP teachers. Check out the material on the courses he is teaching.

Intro to Natural Language Processingon Coursera by U of Michigan

Intro to Artificial Intelligencecourse on Udacity which also covers NLP

Deep Learning for Natural Language Processing (2015 classes)by Richard Socher

Deep Learning for Natural Language Processing (2016 classes)by Richard Socher. Updated to make use of Tensorflow. Note that there are some lectures missing (lecture 9, and lectures 12 onwards).

Natural Language Processing– course on Coursera that was only done in 2013. The videos are not available at the moment. Also Mike Collins is a great professor and his notes and lectures are very good.

Statistical Machine Translation– a Machine Translation course with great assignments and slides.

Natural Language Processing SFU– course byProf Anoop Sarkaron Natural Language Processing. Good notes and some good lectures on youtube about HMM.

Udacity Deep LearningDeep Learning course on Udacity (using Tensorflow) which covers a section on using deep learning for NLP tasks (covering Word2Vec, RNN’s and LSTMs).

NLTK with Python 3 for Natural Language Processingby Harrison Kinsley(sentdex). Good tutorials with NLTK code implementation.

People

Mikolovet al. 2013. Performs well on word similarity and analogy task. Expands on famous example: King – Man + Woman = Queen

Yoav Goldberg

Quoc V. Le

Source Codes

General Machine Learning Topics

Games

Halite: A.I. Coding Game

Vindinium: A.I. Programming Challenge

General Video Game AI Competition

Angry Birds AI Competition

The AI Games

Fighting Game AI Competition

CodeCup

Student StarCraft AI Tournament

AIIDE StarCraft AI Competition

CIG StarCraft AI Competition

CodinGame – AI Bot Games

NLP

Word2Vec source code

FastText Code

gensim

Memory networks are implemented inMemNN. Attempts to solve task of reason attention and memory.

Stack RNN source codeandblog post

Pre-trained word embeddings for WSJ corpusby Koc AI-Lab

HLBL language modelby Turian

Real-valued vector “embeddings”by Dhillon

Improving Word Representations Via Global Context And Multiple Word Prototypesby Huang

Dependency based word embeddings

Global Vectors for Word Representations

TwitIE: An Open-Source Information Extraction Pipeline for Microblog Text

*Node.js and Javascript– – Node.js Libaries for NLP

Twitter-text– A JavaScript implementation of Twitter’s text processing library

Knwl.js– A Natural Language Processor in JS

Retext– Extensible system for analyzing and manipulating natural language

NLP Compromise– Natural Language processing in the browser

Natural– general natural language facilities for node

Python– Python NLP Libraries

Scikit-learn: Machine learning in Python

Natural Language Toolkit (NLTK)

Pattern– A web mining module for the Python programming language. It has tools for natural language processing, machine learning, among others.

TextBlob– Providing a consistent API for diving into common natural language processing (NLP) tasks. Stands on the giant shoulders of NLTK and Pattern, and plays nicely with both.

YAlign– A sentence aligner, a friendly tool for extracting parallel sentences from comparable corpora.

jieba– Chinese Words Segmentation Utilities.

SnowNLP– A library for processing Chinese text.

KoNLPy– A Python package for Korean natural language processing.

Rosetta– Text processing tools and wrappers (e.g. Vowpal Wabbit)

BLLIP Parser– Python bindings for the BLLIP Natural Language Parser (also known as the Charniak-Johnson parser)

PyNLPl– Python Natural Language Processing Library. General purpose NLP library for Python. Also contains some specific modules for parsing common NLP formats, most notably forFoLiA, but also ARPA language models, Moses phrasetables, GIZA++ alignments.

python-ucto– Python binding to ucto (a unicode-aware rule-based tokenizer for various languages)

python-frog– Python binding to Frog, an NLP suite for Dutch. (pos tagging, lemmatisation, dependency parsing, NER)

python-zpar– Python bindings forZPar, a statistical part-of-speech-tagger, constiuency parser, and dependency parser for English.

colibri-core– Python binding to C++ library for extracting and working with with basic linguistic constructions such as n-grams and skipgrams in a quick and memory-efficient way.

spaCy– Industrial strength NLP with Python and Cython.

PyStanfordDependencies– Python interface for converting Penn Treebank trees to Stanford Dependencies.

*C++– – C++ Libraries

MIT Information Extraction Toolkit– C, C++, and Python tools for named entity recognition and relation extraction

CRF++– Open source implementation of Conditional Random Fields (CRFs) for segmenting/labeling sequential data & other Natural Language Processing tasks.

CRFsuite– CRFsuite is an implementation of Conditional Random Fields (CRFs) for labeling sequential data.

BLLIP Parser– BLLIP Natural Language Parser (also known as the Charniak-Johnson parser)

colibri-core– C++ library, command line tools, and Python binding for extracting and working with basic linguistic constructions such as n-grams and skipgrams in a quick and memory-efficient way.

ucto– Unicode-aware regular-expression based tokenizer for various languages. Tool and C++ library. Supports FoLiA format.

libfolia– C++ library for theFoLiA format

frog– Memory-based NLP suite developed for Dutch: PoS tagger, lemmatiser, dependency parser, NER, shallow parser, morphological analyzer.

MeTAMeTA : ModErn Text Analysisis a C++ Data Sciences Toolkit that facilitates mining big text data.

Mecab (Japanese)

Mecab (Korean)

Moses

*Java– – Java NLP Libraries

Stanford NLP

OpenNLP

ClearNLP

Word2vec in Java

ReVerbWeb-Scale Open Information Extraction

OpenRegexAn efficient and flexible token-based regular expression language and engine.

CogcompNLP– Core libraries developed in the U of Illinois’ Cognitive Computation Group.

*Scala– – Scala NLP Libraries

Saul– Library for developing NLP systems, including built in modules like SRL, POS, etc.

Clojure

Clojure-openNLP– Natural Language Processing in Clojure (opennlp)

Infections-clj– Rails-like inflection library for Clojure and ClojureScript

Ruby

Kevin Dias’sA collection of Natural Language Processing (NLP) Ruby libraries, tools and software

Service

Wit-ai– Natural Language Interface for apps and devices.

Iris– Free text search API over large public document collections.

Books

Beginner Books

Data Smart: Using Data Science to Transform Information into Insight 1st Edition

Data Science for Business: What you need to know about data mining and data­ analytic-thinking

Predictive Analytics: The Power to Predict Who Will Click, Buy, Lie, or Die

Practical Books

Machine Learning for Hackers

GitHub repository(R)

GitHub repository(Python)

Python Machine Learning

GitHub repository

Programming Collective Intelligence: Building Smart Web 2.0 Applications

Machine Learning: An Algorithmic Perspective, Second Edition

GitHub repository

Resource repository

Introduction to Machine Learning with Python: A Guide for Data Scientists

GitHub repository

Data Mining: Practical Machine Learning Tools and Techniques, Third Edition

Teaching material

Slides for Chapters 1-5 (zip)

Slides for Chapters 6-8 (zip)

Machine Learning in Action

GitHub repository

Reactive Machine Learning Systems(MEAP)

GitHub repository

An Introduction to Statistical Learning

GitHub repository(R)

GitHub repository(Python)

Videos

Building Machine Learning Systems with Python

GitHub repository

Learning scikit-learn: Machine Learning in Python

GitHub repository

Probabilistic Programming & Bayesian Methods for Hackers

Probabilistic Graphical Models: Principles and Techniques

Machine Learning: Hands-On for Developers and Technical Professionals

Machine Learning Hands-On for Developers and Technical Professionals review

GitHub repository

Learning from Data

Online tutorials

Reinforcement Learning: An Introduction (2nd Edition)

GitHub repository

Machine Learning with TensorFlow(MEAP)

Papers

General Machine Learning Topics

Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning [arXiv]

Overcoming catastrophic forgetting in neural networks [arXiv]

Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer[OpenReview]

A Way out of the Odyssey: Analyzing and Combining Recent Insights for LSTMs [arXiv]

Importance Sampling with Unequal Support [arXiv]

Quasi-Recurrent Neural Networks [arXiv]

Capacity and Learnability in Recurrent Neural Networks [OpenReview]

Unrolled Generative Adversarial Networks [OpenReview]

Deep Information Propagation [OpenReview]

Structured Attention Networks [OpenReview]

Incremental Sequence Learning [arXiv]

b-GAN: Unified Framework of Generative Adversarial Networks [OpenReview]

A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks [OpenReview]

Categorical Reparameterization with Gumbel-Softmax [arXiv]

Computer Vision

Image-to-Image Translation with Conditional Adversarial Networks [arXiv]

Lip Reading Sentences in the Wild [arXiv]

Deep Residual Learning for Image Recognition[arXiv]

Rethinking the Inception Architecture for Computer Vision [arXiv]

Inside-Outside Net: Detecting Objects in Context with Skip Pooling and Recurrent Neural Networks [arXiv]

Deep Speech 2: End-to-End Speech Recognition in English and Mandarin [arXiv]

Reinforcement Learning

Learning to reinforcement learn [arXiv]

Learning to reinforcement learn [arXiv]

A Connection between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy-Based Models [arXiv]

The Predictron: End-To-End Learning and Planning [OpenReview]

Third-Person Imitation Learning[OpenReview]

Generalizing Skills with Semi-Supervised Reinforcement Learning [OpenReview]

Sample Efficient Actor-Critic with Experience Replay [OpenReview]

Reinforcement Learning with Unsupervised Auxiliary Tasks[arXiv]

Neural Architecture Search with Reinforcement Learning [OpenReview]

Towards Information-Seeking Agents [OpenReview]

Multi-Agent Cooperation and the Emergence of (Natural) Language [OpenReview]

Improving Policy Gradient by Exploring Under-appreciated Rewards [OpenReview]

Stochastic Neural Networks for Hierarchical Reinforcement Learning [OpenReview]

Tuning Recurrent Neural Networks with Reinforcement Learning [OpenReview]

RL^2: Fast Reinforcement Learning via Slow Reinforcement Learning [arXiv]

Learning Invariant Feature Spaces to Transfer Skills with Reinforcement Learning [OpenReview]

Learning to Perform Physics Experiments via Deep Reinforcement Learning [OpenReview]

Reinforcement Learning through Asynchronous Advantage Actor-Critic on a GPU [OpenReview]

Learning to Compose Words into Sentences with Reinforcement Learning[OpenReview]

Deep Reinforcement Learning for Accelerating the Convergence Rate [OpenReview]

#Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning[arXiv]

Learning to Compose Words into Sentences with Reinforcement Learning [OpenReview]

Learning to Navigate in Complex Environments [arXiv]

Unsupervised Perceptual Rewards for Imitation Learning [OpenReview]

Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic [OpenReview]

NLP

General Topics

Strategies for Training Large Vocabulary Neural Language Models[arXiv]

Multilingual Language Processing From Bytes[arXiv]

Learning Document Embeddings by Predicting N-grams for Sentiment Classification of Long Movie Reviews[arXiv]

Target-Dependent Sentiment Classification with Long Short Term Memory[arXiv]

Reading Text in the Wild with Convolutional Neural Networks [arXiv]

Deep Reinforcement Learning with a Natural Language Action Space[arXiv]

Sequence Level Training with Recurrent Neural Networks [arXiv]

Teaching Machines to Read and Comprehend[arxiv]

Semi-supervised Sequence Learning[arXiv]

Multi-task Sequence to Sequence Learning[arXiv]

Alternative structures for character-level RNNs[arXiv]

Larger-Context Language Modeling[arXiv]

A Unified Tagging Solution: Bidirectional LSTM Recurrent Neural Network with Word Embedding[arXiv]

Towards Universal Paraphrastic Sentence Embeddings [arXiv]

BlackOut: Speeding up Recurrent Neural Network Language Models With Very Large Vocabularies [arXiv]

Sequence Level Training with Recurrent Neural Networks [arXiv]

Natural Language Understanding with Distributed Representation [arXiv]

sense2vec – A Fast and Accurate Method for Word Sense Disambiguation In Neural Word Embeddings [arXiv]

LSTM-based Deep Learning Models for non-factoid answer selection [arXiv]

Review Articles

Deep Learning for Web Search and Natural Language Processing

Probabilistic topic models

Natural language processing: an introduction

A unified architecture for natural language processing: Deep neural networks with multitask learning

A Critical Review of Recurrent Neural Networksfor Sequence Learning

Deep parsing in Watson

Online named entity recognition method for microtexts in social networking services: A case study of twitter

Word Vectors

A Primer on Neural Network Models for Natural Language ProcessingYoav Goldberg. October 2015. No new info, 75 page summary of state of the art.

A neural probabilistic language modelBengio 2003. Seminal paper on word vectors.

Efficient Estimation of Word Representations in Vector Space

Distributed Representations of Words and Phrases and their Compositionality

Linguistic Regularities in Continuous Space Word Representations

Enriching Word Vectors with Subword Information

Deep Learning, NLP, and Representations

GloVe: Global vectors for word representationPennington, Socher, Manning. 2014. Creates word vectors and relates word2vec to matrix factorizations.Evalutaion section led to controversybyYoav Goldberg

Infinite Dimensional Word Embeddings– new

Skip Thought Vectors– word representation method

Adaptive skip-gram– similar approach, with adaptive properties

Named Entity Recognition

A survey of named entity recognition and classification

Benchmarking the extraction and disambiguation of named entities on the semantic web

Knowledge base population: Successful approaches and challenges

SpeedRead: A fast named entity recognition Pipeline

Sentiment Analysis

Recursive Deep Models for Semantic Compositionality Over a Sentiment TreebankSocher et al. 2013. Introduces Recursive Neural Tensor Network and dataset: “sentiment treebank.” Includesdemo site. Uses a parse tree.

Distributed Representations of Sentences and Documents

Deep Recursive Neural Networks for Compositionality in Language

Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks

Semi-supervised Sequence Learning

Bag of Tricks for Efficient Text Classification

Adversarial Training Methods for Semi-Supervised Text Classification[arXiv]

Neural Machine Translation & Dialog

Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation(abstract)

On Using Very Large Target Vocabulary for Neural Machine Translation

Sequence to Sequence Learning with Neural Networks(nips presentation). Uses seq2seq to generate translations.

Addressing the Rare Word Problem in Neural Machine Translation(abstract)

Effective Approaches to Attention-based Neural Machine Translation

Context-Dependent Word Representation for Neural Machine Translation

Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

Neural Machine Translation by jointly learning to align and translateBahdanau, Cho 2014. “comparable to the existing state-of-the-art phrase-based system on the task of English-to-French translation.” Implements attention mechanism.English to French Demo

Cross-lingual Pseudo-Projected Expectation Regularization for Weakly Supervised Learning

Generating Chinese Named Entity Data from a Parallel Corpus

IXA pipeline: Efficient and Ready to Use Multilingual NLP tools

Iterative Refinement for Machine Translation [OpenReview]

A Convolutional Encoder Model for Neural Machine Translation [arXiv]

Improving Neural Language Models with a Continuous Cache [OpenReview]

Vocabulary Selection Strategies for Neural Machine Translation [OpenReview]

Towards an automatic Turing test: Learning to evaluate dialogue responses [OpenReview]

Dialogue Learning With Human-in-the-Loop [OpenReview]

Batch Policy Gradient Methods for Improving Neural Conversation Models [OpenReview]

Learning through Dialogue Interactions [OpenReview]

Dual Learning for Machine Translation[arXiv]

Unsupervised Pretraining for Sequence to Sequence Learning [arXiv]

Neural Responding Machine for Short-Text ConversationShang et al. 2015 Uses Neural Responding Machine. Trained on Weibo dataset. Achieves one round conversations with 75% appropriate responses.

A Neural Network Approach to Context-Sensitive Generation of Conversational ResponsesSordoni et al. 2015. Generates responses to tweets. UsesRecurrent Neural Network Language Model (RLM) architecture of (Mikolov et al., 2010).source code:RNNLM Toolkit

Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network ModelsSerban, Sordoni, Bengio et al. 2015. Extendshierarchical recurrent encoder-decoderneural network (HRED).

Attention with Intention for a Neural Network Conversation ModelYao et al. 2015 Architecture is three recurrent networks: an encoder, an intention network and a decoder.

A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues

A Neural Conversation ModelVinyals,Le2015. Uses LSTM RNNs to generate conversational responses. Usesseq2seq framework. Seq2Seq was originally designed for machine translation and it “translates” a single sentence, up to around 79 words, to a single sentence response, and has no memory of previous dialog exchanges. Used in GoogleSmart Reply feature for Inbox

Incorporating Copying Mechanism in Sequence-to-Sequence LearningGu et al. 2016 Proposes CopyNet, builds on seq2seq.

A Persona-Based Neural Conversation ModelLi et al. 2016 Proposes persona-based models for handling the issue of speaker consistency in neural response generation. Builds on seq2seq.

Deep Reinforcement Learning for Dialogue GenerationLi et al. 2016. Uses reinforcement learing to generate diverse responses. Trains 2 agents to chat with each other. Builds on seq2seq.

Deep learning for chatbotsArticle summary of state of the art, and challenges for chatbots.

Deep learning for chatbots. part 2Implements a retrieval based dialog agent using dual encoder lstm with TensorFlow, based on the Ubuntu dataset [paper] includessource code

UsesRecurrent Neural Network Language Model (RLM) architecture of (Mikolov et al., 2010).source code:RNNLM Toolkit

Image Captioning

Show, Attend and Tell: Neural Image Caption Generation with Visual AttentionXu et al. 2015 Creates captions by feeding image into a CNN which feeds into hidden state of an RNN that generates the caption. At each time step the RNN outputs next word and the next location to pay attention to via a probability over grid locations. Uses 2 types of attention soft and hard. Soft attention uses gradient descent and backprop and is deterministic. Hard attention selects the element with highest probability. Hard attention uses reinforcement learning, rather than backprop and is stochastic.

Open source implementation in TensorFlow

Memory and Attention Models

Memory Networks

End-To-End Memory NetworksSukhbaatar et. al 2015.

Towards AI-Complete Question Answering: A Set of Prerequisite Toy TasksWeston 2015. Classifies QA tasks like single factoid, yes/no etc. Extends memory networks.

Evaluating prerequisite qualities for learning end to end dialog systemsDodge et. al 2015. Tests Memory Networks on 4 tasks including reddit dialog task. SeeJason Weston lecture on MemNN

Neural Turing Machines

Olah and Carter blog on NTM

Inferring Algorithmic Patterns with Stack-Augmented Recurrent Nets

Reasoning, Attention and Memory RAM workshop at NIPS 2015. slides included

General NLP topics

Neural autocoder for paragraphs and documents– LSTM representation

LSTM over tree structures

Sequence to Sequence Learning– word vectors for machine translation

Teaching Machines to Read and Comprehend– DeepMind paper

Efficient Estimation of Word Representations in Vector Space

Improving distributional similarity with lessons learned from word embeddings

Low-Dimensional Embeddings of Logic

Tutorial on Markov Logic Networks (based on this paper)

Markov Logic Networks for Natural Language Question Answering

Distant Supervision for Cancer Pathway Extraction From Text

Privee: An Architecture for Automatically Analyzing Web Privacy Policies

A Neural Probabilistic Language Model

Template-Based Information Extraction without the Templates

Retrofitting word vectors to semantic lexicons

Unsupervised Learning of the Morphology of a Natural Language

Natural Language Processing (Almost) from Scratch

Computational Grounded Cognition: a new alliance between grounded cognition and computational modelling

Learning the Structure of Biomedical Relation Extractions

Relation extraction with matrix factorization and universal schemas

The Unreasonable Effectiveness of Recurrent Neural Networks

Statistical Language Models based on Neural Networks

Slides from Google Talk

Communities

Quora

Machine Learning

Statistics

Data Mining

Reddit

Machine Learning

Computer Vision

Natural Language

Data Science

Big Data

Statistics

Data Tau

Deep Learning News

KDnuggets

    原文作者:戴德曼
    原文地址: https://www.jianshu.com/p/acf8fd85e622#comments
    本文转自网络文章,转载此文章仅为分享知识,如有侵权,请联系博主进行删除。
点赞