[TOC]

  1. Title: Extracting Action Sequences from Texts Based on Deep Reinforcement Learning
  2. Author: Wenfeng Feng et. al.
  3. Publish Year: Mar 2018
  4. Review Date: Mar 2022

Summary of paper

Motivation

the author want to build a model that learns to directly extract action sequences without external tools like POS tagging and dependency parsing results…

Annotation dataset structure

image-20220315161910319

example

image-20220315162057150

Model

they exploit the framework to learn two models to predict action names and arguments respectively.

image-20220315163000836

image-20220315162925065

Why this approach can be treated as Reinforcement Learning problem

image-20220315174921040

Initially we have word vector and operations pairs with operations all being NULL as starting point.

image-20220315175029764

image-20220315175209178

image-20220315175314781

why the author wants to stick this method to RL schema…

Rewards signals

  1. basic rewards signal
    • image-20220315175854561
  2. additional rewards signal
    • image-20220315180016262
    • image-20220315180205173
    • image-20220315180327017
    • this is used as momentum and buffer based on the overview of the training data
      • issue: we don’t have item rate for testing texts because we don’t have labels.

Potential future work

When we extend to unseen corpus, we may use word association to assist to extract action and action arguments.