Person:
Schmaltz, Allen

Loading...
Profile Picture

Email Address

AA Acceptance Date

Birth Date

Research Projects

Organizational Units

Job Title

Last Name

Schmaltz

First Name

Allen

Name

Schmaltz, Allen

Search Results

Now showing 1 - 3 of 3
  • Publication
    Adapting Sequence Models for Sentence Correction
    (Association for Computational Linguistics, 2017) Schmaltz, Allen; Kim, Yoon; Shieber, Stuart; Rush, Alexander Sasha
    In a controlled experiment of sequence-to-sequence approaches for the task of sentence correction, we find that character-based models are generally more effective than word-based models and models that encode subword information via convolutions, and that modeling the output data as a series of diffs improves effectiveness over standard approaches. Our strongest sequence-to-sequence model improves over our strongest phrase-based statistical machine translation model, with access to the same data, by $6 M^2$ (0.5 GLEU) points. Additionally, in the data environment of the standard CoNLL-2014 setup, we demonstrate that modeling (and tuning against) diffs yields similar or better $M^2$ scores with simpler models and/or significantly less data than previous sequence-to-sequence approaches.
  • Thumbnail Image
    Publication
    Word Ordering Without Syntax
    (Association for Computational Linguistics, 2016) Schmaltz, Allen; Rush, Alexander Sasha; Shieber, Stuart
    Recent work on word ordering has argued that syntactic structure is important, or even required, for effectively recovering the order of a sentence. We find that, in fact, an n-gram language model with a simple heuristic gives strong results on this task. Furthermore, we show that a long short-term memory (LSTM) language model is even more effective at recovering order, with our basic model outperforming a state-of-the-art syntactic model by 11.5 BLEU points. Additional data and larger beams yield further gains, at the expense of training and search time.
  • Thumbnail Image
    Publication
    Sentence-level grammatical error identification as sequence-to-sequence correction
    (Association of Computational Linguistics, 2016) Schmaltz, Allen; Kim, Yoon; Rush, Alexander Sasha; Shieber, Stuart
    We demonstrate that an attention-based encoder-decoder model can be used for sentence-level grammatical error identification for the Automated Evaluation of Scientific Writing (AESW) Shared Task 2016. The attention-based encoder-decoder models can be used for the generation of corrections, in addition to error identification, which is of interest for certain end-user applications. We show that a character-based encoder-decoder model is particularly effective, outperforming other results on the AESW Shared Task on its own, and showing gains over a word-based counterpart. Our final model— a combination of three character-based encoder-decoder models, one word-based encoder-decoder model, and a sentence-level CNN—is the highest performing system on the AESW 2016 binary prediction Shared Task.