.<\/span><\/a><\/p>\n\u56fd\u9645\u7ffb\u8bd1\u52a8\u6001<\/p>\n\n\n\n\n\n\n\n\n\n\n<\/section>\n\n\n\nHow to Teach Large Language Models to Translate Through Self-Reflection<\/strong><\/p>\n<\/section>\n\n\n<\/section>\n<\/section>\n<\/section>\n\n\u5982\u4f55\u901a\u8fc7\u81ea\u6211\u53cd\u601d\u6559\u4f1a\u5927\u8bed\u8a00\u6a21\u578b\u7ffb\u8bd1<\/strong><\/p>\n<\/section>\n<\/section>\n<\/section>\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\nIn a June 12, 2024 paper researchers from Tencent AI and the Harbin Institute of Technology introduced\u00a0TasTe, a method for teaching large language models (LLMs) to translate through self-reflection.<\/span><\/p>\n2024\u5e746\u670812\u65e5\uff0c\u817e\u8baf\u4eba\u5de5\u667a\u80fd\u548c\u54c8\u5c14\u6ee8\u5de5\u4e1a\u5927\u5b66\u7684\u7814\u7a76\u4eba\u5458\u5728\u53d1\u8868\u7684\u8bba\u6587\u4e2d\u63d0\u5230\u4e86\u4e00\u79cd\u6559\u5927\u8bed\u8a00\u6a21\u578b\uff08LLMs\uff09\u901a\u8fc7\u81ea\u6211\u53cd\u601d\u8fdb\u884c\u7ffb\u8bd1\u7684\u65b9\u6cd5\uff0c\u5373TasTe<\/span>\u3002<\/p>\n\n
The key idea is to\u00a0enable LLMs to generate preliminary translations (i.e., drafts), self-evaluate their own translations, and make refinements based on the evaluation.<\/span><\/p>\nTasTe\u7684\u4e3b\u8981\u505a\u6cd5\u662f\u8ba9LLMs\u5148\u521d\u6b65\u7ffb\u8bd1\uff08\u5373\u751f\u6210\u8349\u7a3f\uff09 \uff0c\u7136\u540e\u5bf9\u8bd1\u6587\u81ea\u6211\u8bc4\u4f30\uff0c\u5e76\u6839\u636e\u8bc4\u4f30\u5b8c\u5584\u8bd1\u6587\u3002<\/span><\/p>\n\n
The researchers explained that LLMs have shown exceptional performance across various natural language processing tasks, including machine translation (MT). However, their\u00a0translations<\/span>\u00a0still\u00a0do not match the quality of supervised neural machine translation (NMT) systems.<\/span><\/p>\n\u7814\u7a76\u4eba\u5458\u89e3\u91ca\u8bf4\uff0c\u5927\u8bed\u8a00\u6a21\u578b\u5728\u5404\u79cd\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4efb\u52a1\u4e2d\u8868\u73b0\u51fa\u8272\uff0c\u5305\u62ec\u673a\u5668\u7ffb\u8bd1\uff08MT\uff09\u3002\u7136\u800c\uff0c\u4ed6\u4eec\u7684\u7ffb\u8bd1<\/span>\u4ecd\u7136\u65e0\u6cd5\u4e0e\u6709\u76d1\u7763\u6a21\u5f0f\u7684\u795e\u7ecf\u673a\u5668\u7ffb\u8bd1\uff08NMT\uff09\u7cfb\u7edf\u7684\u8d28\u91cf\u76f8\u5339\u914d\u3002<\/span><\/p>\n\n
To address this, the authors proposed the TasTe framework (translating through self-reflection), which\u00a0improves the translation capabilities of LLMs by incorporating a self-reflection process.<\/span><\/p>\n\u9274\u6b64\uff0c\u4f5c\u8005\u63d0\u51faTasTe\uff0c\u610f\u5728\u901a\u8fc7\u52a0\u5165\u81ea\u6211\u53cd\u601d\u63d0\u9ad8LLMs\u7684\u7ffb\u8bd1\u80fd\u529b\u3002<\/span><\/p>\n\n
This process consists of two stages. In the first stage, LLMs are prompted to generate preliminary translations (i.e. drafts) while simultaneously making quality predictions for these translations.<\/span>\u00a0The quality predictions can be in the form of labels like \u201cgood,\u201d \u201cmedium,\u201d and \u201cbad\u201d or scores ranging from 0 to 100. This self-assessment step allows the models to evaluate the quality of their own outputs.<\/p>\n\u81ea\u6211\u53cd\u601d\u5305\u542b\u4e24\u4e2a\u9636\u6bb5\u3002<\/strong>\u7b2c\u4e00\u9636\u6bb5<\/span>\uff0cLLMs\u7ecf\u63d0\u793a\u751f\u6210\u521d\u6b65\u7ffb\u8bd1\uff0c\u540c\u65f6\u5bf9\u8fd9\u4e9b\u7ffb\u8bd1\u8fdb\u884c\u8d28\u91cf\u9884\u6d4b\u3002<\/span>\u8d28\u91cf\u9884\u6d4b\u53ef\u4ee5\u7528\u201c\u597d\u201d\u3001\u201c\u4e2d\u201d\u548c\u201c\u574f\u201d\u7684\u6807\u7b7e\u8fdb\u884c\u6807\u6ce8\uff0c\u6216\u8005\u7ed9\u51fa0\u5230100\u7684\u8bc4\u5206\u3002\u8fd9\u79cd\u81ea\u6211\u8bc4\u4f30\u7684\u6b65\u9aa4\u53ef\u4ee5\u8ba9\u6a21\u578b\u5bf9\u5176\u8bd1\u6587\u8d28\u91cf\u8fdb\u884c\u8bc4\u4ef7\u3002<\/p>\n\n
In the second stage, LLMs refine these preliminary translations based on the quality predictions in the first stage to produce final translations.<\/span>\u00a0According to Xuebo Liu, Assistant Professor at Harbin Institute of Technology, speaking to Slator, low-quality drafts with severe errors undergo extensive modifications, medium-quality drafts with minor errors receive moderate adjustments, and high-quality drafts with minimal or no errors require little to no changes. \u201cBy equipping models to tailor their modifications to the draft quality, we effectively rectify conspicuous errors and prevent the misguidance of error propagation that could otherwise compromise originally accurate translations, thereby safeguarding the overall translation quality,\u201d he added.<\/p>\n\u7b2c\u4e8c\u9636\u6bb5<\/span>\uff0c LLMs\u5728\u521d\u7ffb\u7684\u57fa\u7840\u4e0a\u7cbe\u8fdb\u8bd1\u6587\uff0c\u4ea7\u51fa\u6700\u7ec8\u8bd1\u6587\u3002<\/span>\u54c8\u5c14\u6ee8\u5de5\u4e1a\u5927\u5b66\u52a9\u7406\u6559\u6388\u5218\u96ea\u6ce2\u5728\u63a5\u53d7Slator\u91c7\u8bbf\u65f6\u8868\u793a\uff1a\u8d28\u91cf\u9884\u6d4b\u5dee\uff0c\u9519\u8bef\u591a\u7684\u521d\u7a3f\u4f1a\u88ab\u5927\u5e45\u4fee\u6539\uff1b\u8d28\u91cf\u9884\u6d4b\u4e2d\u7b49\uff0c\u9519\u8bef\u8f83\u8f7b\u7684\u53ea\u9700\u9002\u5f53\u4fee\u6539\uff1b\u9884\u6d4b\u8d28\u91cf\u9ad8\uff0c\u9519\u8bef\u5f88\u5c11\u751a\u81f3\u6ca1\u6709\u7684\uff0c\u51e0\u4e4e\u4e0d\u9700\u8981\u4fee\u6539\u3002\u4ed6\u8865\u5145\u8bf4\uff1a\u201c\u901a\u8fc7\u8ba9\u6a21\u578b\u6839\u636e\u8349\u7a3f\u8d28\u91cf\u8c03\u6574\u4fee\u6539\uff0c\u6211\u4eec\u53ef\u4ee5\u6709\u6548\u5730\u7ea0\u6b63\u660e\u663e\u7684\u9519\u8bef\uff0c\u5e76\u9632\u6b62\u9519\u8bef\u6269\u6563\uff0c\u8fdb\u800c\u4fdd\u62a4\u6574\u4f53\u7684\u7ffb\u8bd1\u8d28\u91cf\u3002\u8fd9\u79cd\u6269\u6563\u5982\u679c\u4e0d\u52a0\u4ee5\u63a7\u5236\uff0c\u53ef\u80fd\u4f1a\u5f71\u54cd\u539f\u672c\u51c6\u786e\u7684\u7ffb\u8bd1\u3002<\/p>\n\n
This entire process can be seen as a form of self-reflection<\/span>, mirroring the common \u201ctry-evaluate-improve\u201d approach humans use when handling complex tasks to execute them more effectively, Liu said.<\/p>\n\u4ed6\u8fd8\u8bf4\uff0c\u6574\u4e2a\u8fc7\u7a0b\u53ef\u4ee5\u770b\u4f5c\u662f\u4e00\u79cd\u81ea\u6211\u53cd\u601d<\/span>\uff0c\u5c31\u50cf\u4eba\u7c7b\u4e3a\u66f4\u6709\u6548\u5904\u7406\u590d\u6742\u4efb\u52a1\u65f6\u7ecf\u5e38\u4f1a\u91c7\u53d6\u201c\u5c1d\u8bd5-\u8bc4\u4f30-\u6539\u8fdb\u201d\u7684\u65b9\u5f0f\u3002<\/p>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n\nAutomatic Post-Editing Tool<\/strong><\/p>\n<\/section>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\u81ea\u52a8\u8bd1\u540e\u7f16\u8f91\u5de5\u5177<\/strong><\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\nThey evaluated TasTe in four language directions (German \u00a0English and Chinese \u00a0English) using the WMT22 benchmark. They found that\u00a0TasTe outperformed existing methods by effectively utilizing the self-assessment to enhance translation quality.<\/span><\/p>\n\u4e3a\u4e86\u5bf9TasTe\u8fdb\u884c\u8bc4\u4f30\uff0c\u4ed6\u4eec\u4f7f\u7528WMT22\u57fa\u51c6\u5728\u56db\u79cd\u8bed\u8a00\u7ffb\u8bd1\u65b9\u5411\u8fdb\u884c\u4e86\u6d4b\u8bd5\uff08\u82f1\u5fb7\u4e92\u8bd1\u548c\u4e2d\u82f1\u4e92\u8bd1\uff09\uff0c\u53d1\u73b0TasTe\u901a\u8fc7\u6709\u6548\u5229\u7528\u81ea\u6211\u8bc4\u4f30\u6765\u63d0\u9ad8\u7ffb\u8bd1\u8d28\u91cf\uff0c\u8d85\u8d8a\u4e86\u73b0\u6709\u7ffb\u8bd1\u65b9\u6cd5\u3002<\/span><\/p>\n\n
Additionally, they tested if this approach could be used to evaluate translations generated by other systems and refine them as an\u00a0automatic post-editing (APE) tool<\/span>. They found that \u201cTasTe can not only serve as an effective inference framework for a single LLM but also as\u00a0an APE tool to enhance translations generated by other translation systems<\/span>.\u201d<\/p>\n\u6b64\u5916\uff0c\u4ed6\u4eec\u6d4b\u8bd5\u4e86\u8fd9\u79cd\u65b9\u6cd5\u662f\u5426\u53ef\u4ee5\u7528\u6765\u8bc4\u4f30\u5176\u4ed6\u7cfb\u7edf\u751f\u6210\u7684\u7ffb\u8bd1\uff0c\u5e76\u5c06\u5176\u4f5c\u4e3a\u81ea\u52a8\u8bd1\u540e\u7f16\u8f91\uff08APE\uff09\u5de5\u5177<\/span>\u8fdb\u884c\u6539\u8fdb\u3002\u4ed6\u4eec\u53d1\u73b0\uff0c \u201cTasTe\u4e0d\u4ec5\u53ef\u4ee5\u4f5c\u4e3a\u5355\u4e2a\u5927\u8bed\u8a00\u6a21\u578b\u7684\u6709\u6548\u63a8\u7406\u6846\u67b6\uff0c\u8fd8\u53ef\u4ee5\u4f5c\u4e3a\u589e\u5f3a\u5176\u4ed6\u7ffb\u8bd1\u7cfb\u7edf\u751f\u6210\u7ffb\u8bd1\u7684\u81ea\u52a8\u8bd1\u540e\u7f16\u8f91\u5de5\u5177<\/span>\u3002\u201d<\/p>\n\n
The authors provide their code and datasets for further research at\u00a0GitHub<\/span>.<\/p>\n\u4f5c\u8005\u5728GitHub<\/span>\u4e0a\u63d0\u4f9b\u4e86\u4ed6\u4eec\u7684\u4ee3\u7801\u548c\u6570\u636e\u96c6\uff0c\u4ee5\u4f9b\u8fdb\u4e00\u6b65\u7814\u7a76\u3002<\/p>\n\n
Authors: Yutong Wang, Jiali Zeng, Xuebo Liu, Fandong Meng, Jie Zhou, Min Zhang<\/p>\n
\u4f5c\u8005\uff1a\u738b\u5b87\u901a\u3001\u66fe\u4f73\u4e3d\u3001\u5218\u96ea\u6ce2\u3001\u5b5f\u51e1\u4e1c\u3001\u5468\u6d01\u3001\u5f20\u654f<\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n\u539f\u6587\u7f51\u5740\uff1a<\/p>\n
How to Teach Large Language Models to Translate Through Self-Reflection – Slator<\/p>\n<\/section>\n<\/section>\n<\/section>\n\n<\/p>\n\n\n\n\n\n\u7279\u522b\u8bf4\u660e\uff1a\u672c\u6587\u5185\u5bb9\u9009\u81easlator\u5b98\u7f51\uff0c\u4ec5\u4f9b\u5b66\u4e60\u4ea4\u6d41\u4f7f\u7528\uff0c\u5982\u6709\u4fb5\u6743\u8bf7\u540e\u53f0\u8054\u7cfb\u5c0f\u7f16\u5220\u9664<\/span>\u3002<\/span><\/p>\n<\/section>\n<\/blockquote>\n\n
– END –<\/span><\/strong><\/span><\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n <\/p>\n<\/section>\n","protected":false},"excerpt":{"rendered":"
\u56fd\u9645\u7ffb\u8bd1\u52a8\u6001\uff5c\u5982\u4f55\u901a\u8fc7\u81ea\u6211\u53cd\u601d\u6559\u4f1a\u5927\u8bed\u8a00\u6a21\u578b\u7ffb\u8bd1\uff1f \u7ffb\u8bd1\u6280\u672f\u6559\u80b2\u4e0e\u7814\u7a76\u00a02024\u5e7408\u670804\u65e5 00:00\u00a0\u9655 […]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[391],"tags":[],"class_list":["post-32985","post","type-post","status-publish","format-standard","hentry","category-391"],"_links":{"self":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/32985","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=32985"}],"version-history":[{"count":1,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/32985\/revisions"}],"predecessor-version":[{"id":32987,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/32985\/revisions\/32987"}],"wp:attachment":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=32985"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=32985"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=32985"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}