Підтримка
www.wikidata.uk-ua.nina.az
Nejronnij mashinnij pereklad NMT riznovid mashinnogo perekladu yakij peredbachaye vikoristannya shtuchnoyi nejronnoyi merezhi dlya prognozuvannya jmovirnosti poslidovnosti sliv zazvichaj shlyahom operuvannya odrazu cilimi rechennyami v odnij integrovanij modeli VlastivostiNMT modeli vikoristovuyut gliboke navchannya ta navchannya oznakam Dlya yih roboti potribna lishe chastka pam yati neobhidnoyi dlya vikoristannya tradicijnih modelej statistichnogo mashinnogo perekladu SMT Krim togo na vidminu vid zvichajnih sistem perekladu usi chastini modeli nejronnogo perekladu navchayutsya razom vid pochatku do kincya end to end sho daye zmogu maksimizuvati produktivnist perekladu IstoriyaUpershe programi glibokogo navchannya pochali zastosovuvati v 1990 h rokah u zadachah iz rozpiznavannya movlennya Persha naukova stattya pro vikoristannya nejronnih merezh u mashinnomu perekladi z yavilasya u 2014 roci a vprodovzh kilkoh nastupnih iz yiyi dopomogoyu vdalosya virishiti chimalo inshih zadach Stanom na 2017 rik ce taki sistemi NMT sistemi z velikim slovnikom Large Vocabulary NMT NMT sistemi z urahuvannyam chastin sliv Subword NMT Bagatomovni NMT sistemi Multilingual NMT NMT sistemi z bagatma dzherelami Multi Source NMT NMT sistemi bez dzherel Zero Resource NMT NMT sistemi na rivni simvoliv Character dec NMT Povnosimvolni NMT sistemi Fully Character NMT NMT sistemi bez pidgotovki Zero Shot NMT Google Dodatok dlya zahoplennya zobrazhen U 2015 roci NMT sistema vpershe z yavilasya na vidkritomu konkursi mashinnogo perekladu OpenMT 15 Na WMT 15 tezh upershe vzyav uchast NMT pretendent nastupnogo roku sered peremozhciv bulo vzhe 90 NMT sistem Z 2017 roku vikoristovuye nejronnij mashinnij pereklad dlya mittyevogo nadannya dostupu do informaciyi vmishenoyi u globalnij patentnij sistemi Sistema rozroblena u spivpraci z Google pracyuye 31 movoyu Stanom na 2018 rik sistema pereklala ponad 9 mln dokumentiv NapracyuvannyaNMT sistema ye rozvitkom statistichnogo pidhodu na rivni fraz yakij pracyuye na osnovi nezalezhnih pidkomponentiv Nejronnij mashinnij pereklad NMT ne ye kardinalnoyu vidmovoyu vid statistichnogo mashinnogo perekladu SMT Jogo osnovnim vihidnim punktom ye vikoristannya vektornogo podannya sliv i vnutrishnih staniv vbudovuvannya embeddings podannya v neperervnomu prostori continuous space representations Struktura NMT modelej prostisha porivnyano zi strukturoyu modelej na osnovi fraz u nij nemaye okremoyi movnoyi modeli modeli perekladu ta modeli perevporyadkuvannya a ye lishe odna model poslidovnostej yaka peredbachaye odne slovo za raz Odnak ce peredbachennya poslidovnosti sliv spirayetsya odrazu na vse vihidne rechennya ta na vsyu vzhe stvorenu cilovu poslidovnist Pershi sprobi modelyuvannya poslidovnosti sliv zazvichaj provodilisya za dopomogoyu rekurentnoyi nejronnoyi merezhi RNN Dvonapravlena rekurentna nejronna merezha tak zvanij koduvalnik encoder vikoristovuyetsya dlya koduvannya vihidnogo rechennya dlya drugoyi RNN vidomogo yak rozkoduvalnik decoder a ta svoyeyu chergoyu vikoristovuyetsya dlya peredbachennya sliv cilovoyu movoyu Pered rekurentnimi nejronnimi merezhami postayut trudnoshi pid chas koduvannya dovgih vhidnih danih v odin vektor Yih mozhna podolati za dopomogoyu mehanizmu uvagi attention mechanism yakij daye dekoduvalniku zmogu zoseredzhuvatisya na riznih chastinah vhidnih danih pid chas generaciyi kozhnogo vihidnogo slova Isnuyut modeli pokrittya coverage models dlya virishennya problem u takih mehanizmah uvagi yaki prizvodyat do generuvannya nadto dovgogo abo nadto korotkogo perekladu napriklad ignoruvannya nayavnoyi informaciyi pro virivnyuvannya Zgortkovi nejronni merezhi convnets desho krashe obroblyayut dovgi neperervni poslidovnosti ale pevnij chas yih ne vikoristovuvali cherez nayavnist kilkoh nedolikiv U 2017 roci ci nedoliki vdalosya uspishno podolati za dopomogoyu mehanizmiv uvagi Dominuyuchoyu arhitekturoyu dlya kilkoh movnih par zalishayetsya model Transformer yaka bazuyetsya na mehanizmi uvagi Na rivnyah samouvagi ciyeyi modeli doslidzhuyutsya zalezhnosti mizh slovami poslidovnosti shlyahom analizu zv yazkiv mizh usima slovami v parnih poslidovnostyah i bezposerednogo modelyuvannya cih zv yazkiv Cej pidhid prostishij nizh mehanizm selekciyi na yakomu pracyuyut rekurentni nejronni merezhi A jogo prostota dala doslidnikam zmogu rozroblyati visokoyakisni modeli perekladu za dopomogoyu modeli Transformer navit v umovah koli kilkist vhidnih danih nevelika Prikladi zastosuvannyaNejronnij pereklad vikoristovuyetsya u servisah perekladu bagatoh kompanij yak ot Google Microsoft Yandeks PROMT Google vikoristovuye osoblivij riznovid nejronnogo mashinnogo perekladu tak zvanij Google Neural Machine Translation GNMT Majkrosoft vikoristovuye podibnu tehnologiyu dlya perekladu movlennya zokrema u Majkrosoft Perekladachi ta Skype Perekladachi Garvardska grupa z obrobki prirodnoyi movi vipustila OpenNMT sistemu nejronnogo mashinnogo perekladu z vidkritim vihidnih kodom U Yandeks Perekladachi vikoristovuyetsya gibridna model svij variant perekladu proponuye i statistichna model i nejromerezha pislya chogo za dopomogoyu tehnologiyi CatBoost yaka pracyuye na osnovi mashinnogo navchannya vibirayetsya krashij z otrimanih rezultativ Proponuvati sistemi na osnovi nejronnih merezh pochali j inshi postachalniki mashinnogo perekladu zokrema Omniscien Technologies ranishe Asia Online KantanMT SDL Globalese Systran tosho DeepL nadaye zagalnu sistemu perekladu iz sistemami shtuchnogo intelektu glibokogo navchannya Div takozhShtuchna nejronna merezha Mashinne navchannya Perekladach Google Nejronnij mashinnij pereklad Google Microsoft TranslatorPrimitkiKalchbrenner Nal Blunsom Philip 2013 Recurrent Continuous Translation Models Proceedings of the Association for Computational Linguistics 1700 1709 Sutskever Ilya Vinyals Oriol Le Quoc Viet 2014 Sequence to sequence learning with neural networks arXiv 1409 3215 cs CL Kyunghyun Cho Bart van Merrienboer Dzmitry Bahdanau Yoshua Bengio 3 September 2014 On the Properties of Neural Machine Translation Encoder Decoder Approaches arXiv 1409 1259 cs CL OpenMT Challenge 2015 NIST angl 11 veresnya 2015 Procitovano 27 lipnya 2022 WMT15 Machine Translate amer Procitovano 27 lipnya 2022 Bojar Ondrej Chatterjee Rajen Federmann Christian Graham Yvette Haddow Barry Huck Matthias Yepes Antonio Jimeno Koehn Philipp Logacheva Varvara Monz Christof Negri Matteo Neveol Aurelie Neves Mariana Popel Martin Post Matt Rubino Raphael Scarton Carolina Specia Lucia Turchi Marco Verspoor Karin Zampieri Marcos 2016 PDF ACL 2016 First Conference on Machine Translation WMT16 The Association for Computational Linguistics 131 198 Arhiv originalu PDF za 27 sichnya 2018 Procitovano 27 sichnya 2018 Neural Machine Translation European Patent Office 16 lipnya 2018 Procitovano 14 chervnya 2021 Wolk Krzysztof Marasek Krzysztof 2015 Neural based Machine Translation for Medical Text Domain Based on European Medicines Agency Leaflet Texts Procedia Computer Science 64 64 2 9 arXiv 1509 08644 Bibcode 2015arXiv150908644W doi 10 1016 j procs 2015 08 456 S2CID 15218663 Dzmitry Bahdanau Cho Kyunghyun Yoshua Bengio 2014 Neural Machine Translation by Jointly Learning to Align and Translate arXiv 1409 0473 cs CL Bahdanau Dzmitry Cho Kyunghyun Bengio Yoshua 1 veresnya 2014 Neural Machine Translation by Jointly Learning to Align and Translate arXiv 1409 0473 cs CL Tu Zhaopeng Lu Zhengdong Liu Yang Liu Xiaohua Li Hang 2016 Modeling Coverage for Neural Machine Translation arXiv 1601 04811 cs CL Coldewey Devin 29 serpnya 2017 DeepL schools other online translators with clever machine learning TechCrunch Procitovano 27 sichnya 2018 Barrault Loic Bojar Ondrej Costa jussa Marta R Federmann Christian Fishel Mark Graham Yvette Haddow Barry Huck Matthias Koehn Philipp August 2019 Findings of the 2019 Conference on Machine Translation WMT19 Proceedings of the Fourth Conference on Machine Translation Volume 2 Shared Task Papers Day 1 Florence Italy Association for Computational Linguistics 1 61 doi 10 18653 v1 W19 5301 OpenNMT Open Source Neural Machine Translation opennmt net Procitovano 27 lipnya 2022 CatBoost state of the art open source gradient boosting library with categorical features support catboost ai angl Procitovano 27 lipnya 2022 Machine Translation Omniscien Technologies amer Procitovano 27 lipnya 2022 SDL Brings Powerful Cloud Based Neural Machine Translation to Global Brands www rws com amer Procitovano 27 lipnya 2022 Horvath Greg 5 veresnya 2017 Globalese 3 0 released Globalese amer Procitovano 27 lipnya 2022 Neural Machine Translation NMT SYSTRAN www systransoft com angl Procitovano 27 lipnya 2022
Топ