| C90-3044&J93-1006 | 研究方法/流程 | Purple | 提出统计方法 | The method proposed there requires a database to be maintained of the syntactic structures of sentences together with the structures of the corresponding translations.This paper proposes a method to solve this problem. |
| J93-1006&1995.tmi-1.28 | 研究方法/流程 | Purple | 句子对齐算法 | Algorithms for subsentential alignment have been developed as well at granularities of the character a parallel bilingual corpus may be aligned to the sentence level with reasonable accuracyThe algorithm appears to converge to the correct sentence alignment in only a few iterations. partial alignment of the word level to induce a maximum likelihood alignment of the sentence level |
| 1995.tmi-1.28&P96-1021 | 研究方法/流程 | Purple | 反转型录语法(ITG) | inversion transduction grammars or ITGsmaking use of an inversion transduction grammar (ITG) formalism that we recently developed for bilingual language modeling |
| P96-1021&C98-2225 | 模型 | Red | 多项式时间算法 | introduced a polynomial-time algorithmA Polynomial-Time Algorithm for Statistical Machine Translation |
| C98-2225&P01-1050 | 模型 | Red | 随机语法通道 | noisy channel frameworkMachine Translation with a Stochastic Grammatical Channel |
| P01-1050&000223079900018 | 模型 | Red | 利用翻译记忆库 | decoder described in starts from a gloss that uses the translations in the translation memory phrases in the translation memory were automatically extracted from the Viterbi alignments produced by Giza tries to improve on the gloss translation by modifying it incrementally used a statistical translation memory of phrases in conjunction with a statistical translation model uses them in conjunction with a word-for-word statistical translation system extracts phrase translations from automatically aligned corporatranslation memory and a statistical-based translation model set of algorithms that enable us to translate natural language sentences by exploiting both a translation memory and a statistical-based translation model automatically derived translation memory can be used within a statistical framework |
| 000223079900018&N03-1017 | 模型 | Red | 联合短语模型 | joint probabilities estimated by proposed a translation model that assumes that lexical correspondences can be established not only at the word level, but at the phrase level as well joint phrase model proposed by introduced a joint-probability model for phrase translation原标题内容 |
| N03-1017&N04-1033 | 模型 | Red | 评估短语模型 | or the maximum phrase length the phrase extraction method the underlying word alignment model various aspects of phrase-based systems are comparedWe propose a new phrase-based translation model and decoding algorithm that enables us to evaluate and compare several, previously proposed phrase-based translation models heuristic learning of phrase translations from word-based alignments learning phrases longer than three words |
| N04-1033&P05-1033 | 模型 | Red | 单调搜索算法 | simple distortion model that reorders phrases independently of their content, or not at allWe describe a highly efficient monotone search algorithm with a complexity linear in the input sentence length. |
| P05-1033&000274500200121 | 模型 | Red | 同步树替换语法 | the rules discussed in this paper are equivalent to productions of synchronous tree substitution grammarsA Hierarchical Phrase-Based Model for Statistical Machine Translation |
| 000274500200121&P07-1089 | 模型 | Red | 句法树到字符串 | linguistically syntax-based translation models induce tree-to-string translation rules from parallel texts with linguistic annotationsScalable Inference and Training of Context-Rich Syntactic Translation Models |
| P07-1089&P08-1064 | 模型 | Red | 捕获非句法短语 | non-syntactic phrase modelingcapturing nonsyntactic phrase pairs by describing the correspondence between multiple parse trees and one string. |
| P08-1064&D08-1024 | 研究方法/流程 | Purple | 软化树约束 | soften the constraints of the input treeautomatically learns aligned tree sequence pairs with mapping probabilities from word-aligned biparsed parallel texts supports multi-level structure reordering of tree typology with larger span |
| D08-1024&N09-1025 | 模型 | Red | MIRA调优算法 | MIRA is preferable to MERTwe can obtain results using MIRA that match or surpass MERT in terms of both translation quality and computational cost |
| N09-1025&D09-1008 | 研究方法/流程 | Purple | 大规模新特征 | small fraction of the bi-lingual features available11,001 New Features for Statistical Machine Translation |
| D09-1008&C10-1050 | 研究方法/流程 | Purple | 集成依赖结构 | proposed a way to integrate dependency structure into target and source side string on hierarchical phrase rulesstate-of-the-art hierarchical MT system source dependency LM scores |
| C10-1050&D11-1079 | 模型 | Red | 基于词的重排序 | take a word-based reordering approach for HPBMTHierarchical Phrase-based Machine Translation with Word-based Reordering Model |
| D11-1079&D13-1053 | 模型 | Red | 软依赖约束重排序 | improving the reordering of hierarchical phrase-based translation systems external reordering modelsWe derive soft constraints from the source dependency parsing to directly address the reordering problem for the hierarchical phrasebased model Soft Dependency Constraints for Reordering in Hierarchical Phrase-Based Translation |
| D13-1053&P14-1129 | 研究方法/流程 | Purple | 分解源句法 | Factored source syntaxFactored Soft Source Syntactic Constraints for Hierarchical Machine Translation |
| P14-1129&Q15-1013 | 模型 | Red | 集成与源上下文 | ensembles of networks outperform a single network source-context as inaugments the NNLM with a source context window |
| Q15-1013&D15-1248 | 模型 | Red | 关系依赖语言模型 | dependency language model (RDLM) relational dependency language model We propose a language model for dependency structures that is relational rather than configurational as a feature function in string-to-tree SMT from English to German and Russian |
| D15-1248&000493806800162 | 模型 | Red | 联合形态句法模型 | syntaxbased SMT resultsimprovements in translation quality of 1.4-1.8 BLEU We propose to model syntactic and morphological structure jointly in a dependency translation model A Joint Dependency Model of Morphological and Syntactic Structure for Statistical Machine Translation |
| 000493806800162&000493806800100 | 研究方法/流程 | Purple | 子词单元处理稀疏词 | this problemNeural Machine Translation of Rare Words with Subword Units |
| 000493806800100&000493805000058 | 模型 | Red | 混合词-字符模型 | combination of word and charactersAchieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models |
| 000493805000058&D16-1026 | 模型 | Red | 字符级NMT | sequences of charactersCharacter-based Neural Machine Translation |
| D16-1026&Q17-1026 | 研究方法/流程 | Purple | 零资源翻译 | each minibatch is balanced the proportion of each language pair in a single minibatch corresponds to that of the full corpusZero-Resource Translation with Multi-Lingual Neural Machine Translation |
| Q17-1026&000493992300092 | 模型 | Red | 全字符级NMT | character level character or sub-word levelsFully Character-Level Neural Machine Translation without Explicit Segmentation |
| 000493992300092&000485488905004 | 模型 | Red | 分层级句法翻译 | syntax informationthe target sentence is translated hierarchically from chunks to words, with information in different granularities being leveraged |
| 000485488905004&000386658300285 | 数据集 | Green | 多领域平行语料 | discourse parallel corpus 500 thousand Portuguese-English sentence pairs in various domains such as news, laws, microblog etc. 2 million Chinese-EnglishExperimental results on both Chinese English |
| 000386658300285&W11-2132 | 研究方法/流程 | Purple | SMT领域自适应 | approachDomain Adaptation for Statistical Machine Translation |
| W11-2132&E12-1014 | 研究方法/流程 | Purple | 利用单语数据 | use an existing SMT system to discover parallel sentences within independent monolingual texts, and use them to re-train and enhance the systemInvestigations on Translation Model Adaptation Using Monolingual Data |
| E12-1014&P13-1036 | 研究方法/流程 | Purple | 单语诱导短语表 | uses monolingual corpora to induce phrase tablesToward Statistical Machine Translation without Parallel Corpora We extend existing research on bilingual lexicon induction to estimate both lexical and phrasal translation probabilities for MT-scale phrasetables. We estimate the parameters of a phrasebased statistical machine translation system from monolingual corpora instead of a bilingual parallel corpus. |
| P13-1036&D14-1061 | 模型 | Red | 贝叶斯哈希采样 | Bayesian inference make decipherment scalablewe propose a new Bayesian inference method Scalable Decipherment for Machine Translation via Hash Sampling |
| D14-1061&000493806800185 | 研究方法/流程 | Purple | 联合对齐与破译 | decipherment problemBeyond Parallel Data: Joint Word Alignment and Decipherment Improves Machine Translation |
| 000493806800185&D16-1160 | 研究方法/流程 | Purple | 半监督学习 | propose a similar semi-supervised framework similar semi-supervised reconstruction methodSemi-Supervised Learning for Neural Machine Translation |
| D16-1160&D16-1249 | 研究方法/流程 | Purple | 利用源端单语数据 | achieve better results monolingual corporaExploiting Source-side Monolingual Data in Neural Machine Translation |
| D16-1249&D16-1096 | 研究方法/流程 | Purple | 监督注意力机制 | supervised alignmentsSupervised Attentions for Neural Machine Translation |
| D16-1096&000493984800177 | 模型 | Red | 覆盖率模型 | coverage modelCoverage Embedding Models for Neural Machine Translation |
| 000493984800177&D17-1012 | 模型 | Red | NMT引入句法 | use of syntactic structures in NMT modelsImproved Neural Machine Translation with a Syntax-Aware Encoder and Decoder |
| D17-1012&D17-1209 | 模型 | Red | 隐式句法图解析 | incorporate syntactic information in NMT models in a relatively indirect way (e.g., multi-task learninglearns a latent graph parser as part of the encoder of an attention-based neural machine translation model jointly learns translation and source-side latent graph representations of sentences |
| D17-1209&Q18-1017 | 研究方法/流程 | Purple | 图卷积编码器 | Providing explicit linguistic informationGraph Convolutional Encoders for Syntax-aware Neural Machine Translation |
| Q18-1017&C18-1263 | 研究方法/流程 | Purple | 调度式多任务学习 | multi-task learning frameworksScheduled Multi-Task Learning: From Syntax to Translation |
| C18-1263&000900116903092 | 模型 | Red | 任务特定注意力 | sharing all parameters but the attention mechanismallowing for language-specific specialization of the attention model to a particular language-pair or task Multilingual Neural Machine Translation with Task-Specific Attention compared to a model in which all parameters are shared |
| 000900116903092&000736531900047 | 模型 | Red | 大规模多对多模型 | multilingual translation jointly train one translation model that translates multiple language directions at the same time shares representations to improve the translation performance on low-resource languagesmassively multilingual many-to-many models are effective in low resource settings Massively Multilingual Neural Machine Translation training a single model that supports translation from multiple source languages into multiple target languages |
| 000736531900047&001181866502027 | 模型 | Red | mBART降噪预训练 | mBART fine-tunes with traditional on-the-fly backtranslation pre-trains on a variety of language configurationsMultilingual Denoising Pre-training for Neural Machine Translation |
| 001181866502027&000570978201083 | 研究方法/流程 | Purple | 统一概率框架 | generative modelingour approach results in higher BLEU scores over state-of-the-art unsupervised models We present a probabilistic framework for multilingual neural machine translation that encompasses supervised and unsupervised setups, focusing on unsupervised translation |
| 000570978201083&000663162000001 | 模型 | Red | 在线回译 | online backtranslation to improve the performance of non-English pairs multilingual translation models factorize computation when translating to many languages and share information between similar languages English-Centric multilingual model from on the OPUS100 corpus Using backtranslation thus requires the ability to translate in both directions Mining data for each and every language pair is prohibitive - previous work circumvents this issue by focusing only on the 99 pairs that go through English train on 100 directions increasing model capacityExperiments on OPUS-100 (a novel multilingual dataset with 100 languages) Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation stronger modeling capacity OPUS-100 (a novel multilingual dataset with 100 languages) propose random online backtranslation |
| 000663162000001&000698663100025 | 模型 | Red | 多对多翻译模型 | multilingual NMT, which supports translation from multiple source languages into multiple target languages with a single modeltraining a single model able to translate between any pair of languages create a true Many-to-Many multilingual translation model that can translate directly between any pair of 100 languages |
| 000698663100025&000860727001027 | 模型 | Red | 语言特定子网络(LaSS) | update part of the MT parametersLearning Language Specific Sub-network (LaSS) for each language pair to counter parameter interference |
| 000860727001027&000847347700003 | 研究方法/流程 | Purple | 检索增强翻译(KSTER) | semi-parametric models retrieve similar examples from this datastorepropose to learn Kernel-Smoothed Translation with Example Retrieval (KSTER) non-parametric approaches that retrieve similar examples from a database to guide the translation process |
| 000847347700003&2023.findings-acl.564 | 研究方法/流程 | Purple | 推理时域外数据 | utilizes out-of-domain bitext during inferenceEfficient Machine Translation Domain Adaptation |
| 2023.findings-acl.564&001371932502043 | 研究方法/流程 | Purple | 上下文示例选择 | the quality of the few-shot demonstration impacts the quality of the output.In-context Examples Selection for Machine Translation |
| 001371932502043&001156229800010 | 模型 | Red | LLM翻译能力评估 | LLMs can produce fluent and adequate translations, especially for high-resource English-centric language pairs, that are competitive with those of dedicated supervised translation models translation errors, even severely critical ones, obtained via prompting a LLM are different from those produced by traditional machine translation modelsa transformer decoder-only model trained solely with self-supervised learning is able to match specialized supervised state-of-the-art models |
| 001156229800010&2023.americasnlp-1.17 | 模型 | Red | LLM幻觉分析 | It remains unclear how well LLMs perform when translating into extremely low-resource languages hallucinations are qualitatively different from conventional translation modelsLLMs produce qualitatively different hallucinations to those of NMT models models struggle with hallucinations primarily in low-resource directions |
| 2023.americasnlp-1.17&2023.americasnlp-1.23 | 模型 | Red | kNN-MT | describes systems which make use of kNN and an external data store during decoding submitted four different models for all language pairscombination of fine-tuned M2M100 with $k$NN-MT submitted a total of four systems including GPT-4, a bilingual model, fine-tuned M2M100, and a combination of fine-tuned M2M100 with $k$NN-MT |
| 2023.americasnlp-1.23&2023.americasnlp-1.21 | 研究方法/流程 | Purple | AmericasNLP评测 | The 2023 AmericasNLP Shared TaskFindings of the AmericasNLP 2023 Shared Task on Machine Translation into Indigenous Languages |
| 2023.americasnlp-1.21&2024.findings-acl.670 | 研究方法/流程 | Purple | 单语数据回译 | back-translations produced from monolingual data constitutions, handbooks, news articlesavailable data from various sources such as constitutions, handbooks and news articles |