Table 1 基于拓扑权重的主路径

节点id节点类型节点颜色中文概括知识单元
C90-3044&J93-1006研究方法/流程Purple提出统计方法The method proposed there requires a database to be maintained of the syntactic structures of sentences together with the structures of the corresponding translations.This paper proposes a method to solve this problem.
J93-1006&1995.tmi-1.28研究方法/流程Purple句子对齐算法Algorithms for subsentential alignment have been developed as well at granularities of the character a parallel bilingual corpus may be aligned to the sentence level with reasonable accuracyThe algorithm appears to converge to the correct sentence alignment in only a few iterations. partial alignment of the word level to induce a maximum likelihood alignment of the sentence level
1995.tmi-1.28&P96-1021研究方法/流程Purple反转型录语法(ITG)inversion transduction grammars or ITGsmaking use of an inversion transduction grammar (ITG) formalism that we recently developed for bilingual language modeling
P96-1021&C98-2225模型Red多项式时间算法introduced a polynomial-time algorithmA Polynomial-Time Algorithm for Statistical Machine Translation
C98-2225&P01-1050模型Red随机语法通道noisy channel frameworkMachine Translation with a Stochastic Grammatical Channel
P01-1050&000223079900018模型Red利用翻译记忆库decoder described in starts from a gloss that uses the translations in the translation memory phrases in the translation memory were automatically extracted from the Viterbi alignments produced by Giza tries to improve on the gloss translation by modifying it incrementally used a statistical translation memory of phrases in conjunction with a statistical translation model uses them in conjunction with a word-for-word statistical translation system extracts phrase translations from automatically aligned corporatranslation memory and a statistical-based translation model set of algorithms that enable us to translate natural language sentences by exploiting both a translation memory and a statistical-based translation model automatically derived translation memory can be used within a statistical framework
000223079900018&N03-1017模型Red联合短语模型joint probabilities estimated by proposed a translation model that assumes that lexical correspondences can be established not only at the word level, but at the phrase level as well joint phrase model proposed by introduced a joint-probability model for phrase translation原标题内容
N03-1017&N04-1033模型Red评估短语模型or the maximum phrase length the phrase extraction method the underlying word alignment model various aspects of phrase-based systems are comparedWe propose a new phrase-based translation model and decoding algorithm that enables us to evaluate and compare several, previously proposed phrase-based translation models heuristic learning of phrase translations from word-based alignments learning phrases longer than three words
N04-1033&P05-1033模型Red单调搜索算法simple distortion model that reorders phrases independently of their content, or not at allWe describe a highly efficient monotone search algorithm with a complexity linear in the input sentence length.
P05-1033&000274500200121模型Red同步树替换语法the rules discussed in this paper are equivalent to productions of synchronous tree substitution grammarsA Hierarchical Phrase-Based Model for Statistical Machine Translation
000274500200121&P07-1089模型Red句法树到字符串linguistically syntax-based translation models induce tree-to-string translation rules from parallel texts with linguistic annotationsScalable Inference and Training of Context-Rich Syntactic Translation Models
P07-1089&P08-1064模型Red捕获非句法短语non-syntactic phrase modelingcapturing nonsyntactic phrase pairs by describing the correspondence between multiple parse trees and one string.
P08-1064&D08-1024研究方法/流程Purple软化树约束soften the constraints of the input treeautomatically learns aligned tree sequence pairs with mapping probabilities from word-aligned biparsed parallel texts supports multi-level structure reordering of tree typology with larger span
D08-1024&N09-1025模型RedMIRA调优算法MIRA is preferable to MERTwe can obtain results using MIRA that match or surpass MERT in terms of both translation quality and computational cost
N09-1025&D09-1008研究方法/流程Purple大规模新特征small fraction of the bi-lingual features available11,001 New Features for Statistical Machine Translation
D09-1008&C10-1050研究方法/流程Purple集成依赖结构proposed a way to integrate dependency structure into target and source side string on hierarchical phrase rulesstate-of-the-art hierarchical MT system source dependency LM scores
C10-1050&D11-1079模型Red基于词的重排序take a word-based reordering approach for HPBMTHierarchical Phrase-based Machine Translation with Word-based Reordering Model
D11-1079&D13-1053模型Red软依赖约束重排序improving the reordering of hierarchical phrase-based translation systems external reordering modelsWe derive soft constraints from the source dependency parsing to directly address the reordering problem for the hierarchical phrasebased model Soft Dependency Constraints for Reordering in Hierarchical Phrase-Based Translation
D13-1053&P14-1129研究方法/流程Purple分解源句法Factored source syntaxFactored Soft Source Syntactic Constraints for Hierarchical Machine Translation
P14-1129&Q15-1013模型Red集成与源上下文ensembles of networks outperform a single network source-context as inaugments the NNLM with a source context window
Q15-1013&D15-1248模型Red关系依赖语言模型dependency language model (RDLM) relational dependency language model We propose a language model for dependency structures that is relational rather than configurational as a feature function in string-to-tree SMT from English to German and Russian
D15-1248&000493806800162模型Red联合形态句法模型syntaxbased SMT resultsimprovements in translation quality of 1.4-1.8 BLEU We propose to model syntactic and morphological structure jointly in a dependency translation model A Joint Dependency Model of Morphological and Syntactic Structure for Statistical Machine Translation
000493806800162&000493806800100研究方法/流程Purple子词单元处理稀疏词this problemNeural Machine Translation of Rare Words with Subword Units
000493806800100&000493805000058模型Red混合词-字符模型combination of word and charactersAchieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models
000493805000058&D16-1026模型Red字符级NMTsequences of charactersCharacter-based Neural Machine Translation
D16-1026&Q17-1026研究方法/流程Purple零资源翻译each minibatch is balanced the proportion of each language pair in a single minibatch corresponds to that of the full corpusZero-Resource Translation with Multi-Lingual Neural Machine Translation
Q17-1026&000493992300092模型Red全字符级NMTcharacter level character or sub-word levelsFully Character-Level Neural Machine Translation without Explicit Segmentation
000493992300092&000485488905004模型Red分层级句法翻译syntax informationthe target sentence is translated hierarchically from chunks to words, with information in different granularities being leveraged
000485488905004&000386658300285数据集Green多领域平行语料discourse parallel corpus 500 thousand Portuguese-English sentence pairs in various domains such as news, laws, microblog etc. 2 million Chinese-EnglishExperimental results on both Chinese English
000386658300285&W11-2132研究方法/流程PurpleSMT领域自适应approachDomain Adaptation for Statistical Machine Translation
W11-2132&E12-1014研究方法/流程Purple利用单语数据use an existing SMT system to discover parallel sentences within independent monolingual texts, and use them to re-train and enhance the systemInvestigations on Translation Model Adaptation Using Monolingual Data
E12-1014&P13-1036研究方法/流程Purple单语诱导短语表uses monolingual corpora to induce phrase tablesToward Statistical Machine Translation without Parallel Corpora We extend existing research on bilingual lexicon induction to estimate both lexical and phrasal translation probabilities for MT-scale phrasetables. We estimate the parameters of a phrasebased statistical machine translation system from monolingual corpora instead of a bilingual parallel corpus.
P13-1036&D14-1061模型Red贝叶斯哈希采样Bayesian inference make decipherment scalablewe propose a new Bayesian inference method Scalable Decipherment for Machine Translation via Hash Sampling
D14-1061&000493806800185研究方法/流程Purple联合对齐与破译decipherment problemBeyond Parallel Data: Joint Word Alignment and Decipherment Improves Machine Translation
000493806800185&D16-1160研究方法/流程Purple半监督学习propose a similar semi-supervised framework similar semi-supervised reconstruction methodSemi-Supervised Learning for Neural Machine Translation
D16-1160&D16-1249研究方法/流程Purple利用源端单语数据achieve better results monolingual corporaExploiting Source-side Monolingual Data in Neural Machine Translation
D16-1249&D16-1096研究方法/流程Purple监督注意力机制supervised alignmentsSupervised Attentions for Neural Machine Translation
D16-1096&000493984800177模型Red覆盖率模型coverage modelCoverage Embedding Models for Neural Machine Translation
000493984800177&D17-1012模型RedNMT引入句法use of syntactic structures in NMT modelsImproved Neural Machine Translation with a Syntax-Aware Encoder and Decoder
D17-1012&D17-1209模型Red隐式句法图解析incorporate syntactic information in NMT models in a relatively indirect way (e.g., multi-task learninglearns a latent graph parser as part of the encoder of an attention-based neural machine translation model jointly learns translation and source-side latent graph representations of sentences
D17-1209&Q18-1017研究方法/流程Purple图卷积编码器Providing explicit linguistic informationGraph Convolutional Encoders for Syntax-aware Neural Machine Translation
Q18-1017&C18-1263研究方法/流程Purple调度式多任务学习multi-task learning frameworksScheduled Multi-Task Learning: From Syntax to Translation
C18-1263&000900116903092模型Red任务特定注意力sharing all parameters but the attention mechanismallowing for language-specific specialization of the attention model to a particular language-pair or task Multilingual Neural Machine Translation with Task-Specific Attention compared to a model in which all parameters are shared
000900116903092&000736531900047模型Red大规模多对多模型multilingual translation jointly train one translation model that translates multiple language directions at the same time shares representations to improve the translation performance on low-resource languagesmassively multilingual many-to-many models are effective in low resource settings Massively Multilingual Neural Machine Translation training a single model that supports translation from multiple source languages into multiple target languages
000736531900047&001181866502027模型RedmBART降噪预训练mBART fine-tunes with traditional on-the-fly backtranslation pre-trains on a variety of language configurationsMultilingual Denoising Pre-training for Neural Machine Translation
001181866502027&000570978201083研究方法/流程Purple统一概率框架generative modelingour approach results in higher BLEU scores over state-of-the-art unsupervised models We present a probabilistic framework for multilingual neural machine translation that encompasses supervised and unsupervised setups, focusing on unsupervised translation
000570978201083&000663162000001模型Red在线回译online backtranslation to improve the performance of non-English pairs multilingual translation models factorize computation when translating to many languages and share information between similar languages English-Centric multilingual model from on the OPUS100 corpus Using backtranslation thus requires the ability to translate in both directions Mining data for each and every language pair is prohibitive - previous work circumvents this issue by focusing only on the 99 pairs that go through English train on 100 directions increasing model capacityExperiments on OPUS-100 (a novel multilingual dataset with 100 languages) Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation stronger modeling capacity OPUS-100 (a novel multilingual dataset with 100 languages) propose random online backtranslation
000663162000001&000698663100025模型Red多对多翻译模型multilingual NMT, which supports translation from multiple source languages into multiple target languages with a single modeltraining a single model able to translate between any pair of languages create a true Many-to-Many multilingual translation model that can translate directly between any pair of 100 languages
000698663100025&000860727001027模型Red语言特定子网络(LaSS)update part of the MT parametersLearning Language Specific Sub-network (LaSS) for each language pair to counter parameter interference
000860727001027&000847347700003研究方法/流程Purple检索增强翻译(KSTER)semi-parametric models retrieve similar examples from this datastorepropose to learn Kernel-Smoothed Translation with Example Retrieval (KSTER) non-parametric approaches that retrieve similar examples from a database to guide the translation process
000847347700003&2023.findings-acl.564研究方法/流程Purple推理时域外数据utilizes out-of-domain bitext during inferenceEfficient Machine Translation Domain Adaptation
2023.findings-acl.564&001371932502043研究方法/流程Purple上下文示例选择the quality of the few-shot demonstration impacts the quality of the output.In-context Examples Selection for Machine Translation
001371932502043&001156229800010模型RedLLM翻译能力评估LLMs can produce fluent and adequate translations, especially for high-resource English-centric language pairs, that are competitive with those of dedicated supervised translation models translation errors, even severely critical ones, obtained via prompting a LLM are different from those produced by traditional machine translation modelsa transformer decoder-only model trained solely with self-supervised learning is able to match specialized supervised state-of-the-art models
001156229800010&2023.americasnlp-1.17模型RedLLM幻觉分析It remains unclear how well LLMs perform when translating into extremely low-resource languages hallucinations are qualitatively different from conventional translation modelsLLMs produce qualitatively different hallucinations to those of NMT models models struggle with hallucinations primarily in low-resource directions
2023.americasnlp-1.17&2023.americasnlp-1.23模型RedkNN-MTdescribes systems which make use of kNN and an external data store during decoding submitted four different models for all language pairscombination of fine-tuned M2M100 with $k$NN-MT submitted a total of four systems including GPT-4, a bilingual model, fine-tuned M2M100, and a combination of fine-tuned M2M100 with $k$NN-MT
2023.americasnlp-1.23&2023.americasnlp-1.21研究方法/流程PurpleAmericasNLP评测The 2023 AmericasNLP Shared TaskFindings of the AmericasNLP 2023 Shared Task on Machine Translation into Indigenous Languages
2023.americasnlp-1.21&2024.findings-acl.670研究方法/流程Purple单语数据回译back-translations produced from monolingual data constitutions, handbooks, news articlesavailable data from various sources such as constitutions, handbooks and news articles