\n01<\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n
\n\n\n\n<\/section>\n<\/section>\n\n\n
2024\u5e749\u67086\u65e5,\u6765\u81ea\u666e\u7eb3\u8ba1\u7b97\u673a\u6280\u672f\u7814\u7a76\u6240\u7684\u7279\u8d3e\u65af\u00b7\u5fb7\u4ec0\u6f58\u5fb7(Tejas Deshpande)\u548c\u5c3c\u8fea\u00b7\u79d1\u74e6\u5854\u5c14(Nidhi Kowtal),\u4ee5\u53ca\u5370\u5ea6\u7406\u5de5\u5b66\u9662\u9a6c\u5fb7\u62c9\u65af\u5206\u6821\u7684\u62c9\u7ef4\u62c9\u6770\u00b7\u4e54\u5e0c(Raviraj Joshi)\u5728\u4e00\u7bc7\u8bba\u6587\u4e2d\u63d0\u51fa\u4e86\u94fe\u5f0f\u7ffb\u8bd1\u63d0\u793a(Chain-of-Translation Prompting, CoTR)\u3002\u8fd9\u662f\u4e00\u79cd\u65e8\u5728\u63d0\u5347\u5927\u578b\u8bed\u8a00\u6a21\u578b(LLMs)\u5728\u4f4e\u8d44\u6e90\u8bed\u8a00\u5904\u7406\u80fd\u529b\u7684\u65b0\u578b\u63d0\u793a\u6280\u672f\u3002<\/span><\/strong><\/span><\/p>\n\u00a0<\/span><\/p>\nIn a September 6, 2024,\u00a0paper, Tejas Deshpande and Nidhi Kowtal from the Pune Institute of Computer Technology, along with Raviraj Joshi from the Indian Institute of Technology Madras, introduced Chain-of-Translation Prompting (CoTR),\u00a0a new prompting technique designed to improve the performance of large language models\u00a0(LLMs) for low-resource languages.<\/span><\/strong><\/p>\n\n
\u7814\u7a76\u4eba\u5458\u89e3\u91ca\u9053,\u7531\u4e8e\u8bad\u7ec3\u6216\u5fae\u8c03\u6570\u636e\u6709\u9650,\u591a\u8bed\u8a00LLMs\u5728\u5904\u7406\u4f4e\u8d44\u6e90\u8bed\u8a00\u7684\u8f93\u5165\u53e5\u5b50\u65f6\u5e38\u9047\u5230\u56f0\u96be\u3002<\/strong>\u56e0\u6b64,”\u4f4e\u8d44\u6e90\u8bed\u8a00\u4f7f\u7528\u8005\u5f80\u5f80\u65e0\u6cd5\u4eab\u53d7\u5230\u5148\u8fdb\u81ea\u7136\u8bed\u8a00\u5904\u7406(NLP)\u6280\u672f\u5e26\u6765\u7684\u76ca\u5904,”\u7814\u7a76\u4eba\u5458\u5f3a\u8c03\u4e86\u5f00\u53d1\u65b0\u6280\u672f\u4ee5\u7f29\u5c0f\u8fd9\u4e00\u5dee\u8ddd\u7684\u5fc5\u8981\u6027\u3002<\/p>\n\n
The researchers explained that multilingual LLMs struggle to process input sentences (i.e., the actual text the LLM has to work on) in low-resource languages due to the limited data available for training or fine-tuning.<\/strong>\u00a0As a result, \u201cspeakers of low-resource languages are frequently excluded from the benefits of advanced NLP technologies,\u201d said the researchers, emphasizing the need for new techniques to close this gap.<\/p>\n\n
\u4e3a\u5e94\u5bf9\u8fd9\u4e00\u6311\u6218,\u4ed6\u4eec\u63a2\u7d22\u4e86\u5229\u7528LLMs\u591a\u8bed\u8a00\u7ffb\u8bd1\u80fd\u529b\u7684\u65b0\u578b\u63d0\u793a\u7b56\u7565,\u5e76\u7531\u6b64\u63d0\u51fa\u4e86CoTR\u3002<\/strong><\/span><\/p>\n\n
To address this challenge, they explored new prompting strategies that leverage the multilingual translation abilities of LLMs and introduced CoTR.<\/strong><\/span><\/p>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n\n\n\n02<\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n
\n\n\n\nCoTR\u91cd\u6784\u4e86\u4f20\u7edf\u7684\u63d0\u793a\u65b9\u5f0f:\u9996\u5148\u5c06\u4f4e\u8d44\u6e90\u8bed\u8a00\u7684\u8f93\u5165\u53e5\u5b50\u7ffb\u8bd1\u6210\u8d44\u6e90\u66f4\u4e30\u5bcc\u7684\u8bed\u8a00(\u5982\u82f1\u8bed),LLMs\u901a\u5e38\u5728\u8fd9\u4e9b\u8bed\u8a00\u4e2d\u8868\u73b0\u66f4\u4f73\u3002\u968f\u540e,LLM\u5728\u7ffb\u8bd1\u540e\u7684\u6587\u672c\u4e0a\u6267\u884cNLP\u4efb\u52a1(\u5982\u60c5\u611f\u5206\u6790\u6216\u6587\u672c\u751f\u6210),\u6700\u540e\u53ef\u9009\u62e9\u5c06\u8f93\u51fa\u91cd\u65b0\u7ffb\u8bd1\u56de\u539f\u59cb\u8bed\u8a00\u3002\u7814\u7a76\u4eba\u5458\u5f3a\u8c03,”\u6240\u6709\u8fd9\u4e9b\u6b65\u9aa4\u90fd\u96c6\u6210\u5728\u4e00\u4e2a\u5355\u4e00\u63d0\u793a\u4e2d\u3002”<\/span><\/strong><\/p>\n\n
CoTR restructures the traditional prompts by first translating the input sentence from a low-resource language into a higher-resource language, such as English, where LLMs typically perform better.\u00a0The LLM then executes the NLP task \u2014 such as sentiment analysis or text generation\u00a0\u2014 on the translated text, followed by an optional retranslation of the output back into the original language. \u201cAll these steps are specified in a single prompt,\u201d the researchers emphasized.<\/span><\/strong><\/p>\n\u00a0<\/span><\/strong><\/p>\nCoTR\u9002\u7528\u4e8e\u591a\u79cd\u4efb\u52a1,\u5305\u62ec\u60c5\u611f\u5206\u6790\u3001\u4ec7\u6068\u8a00\u8bba\u5206\u7c7b\u3001\u4e3b\u9898\u5206\u7c7b\u548c\u6587\u672c\u751f\u6210\u7b49\u3002<\/span><\/p>\n\n
CoTR can be applied to various tasks, including sentiment analysis, hate speech classification, subject classification, and text generation.\u00a0<\/span><\/p>\n\n
\u7814\u7a76\u56e2\u961f\u9009\u62e9\u4e86\u9a6c\u62c9\u5730\u8bed\u4f5c\u4e3aCoTR\u7684\u6d4b\u8bd5\u5bf9\u8c61\u3002\u9a6c\u62c9\u5730\u8bed\u662f\u4e00\u79cd\u5370\u5ea6\u8bed\u8a00,\u867d\u6709\u5927\u91cf\u4f7f\u7528\u8005,\u4f46\u6570\u5b57\u548c\u8bed\u8a00\u8d44\u6e90\u532e\u4e4f,\u8fd9\u4f7f\u5f97NLP\u6a21\u578b\u96be\u4ee5\u6709\u6548\u5904\u7406\u3002<\/p>\n
\n
The researchers tested CoTR on Marathi, an Indic language with a significant speaker base but insufficient digital and linguistic resources, making it a challenge for NLP models to handle.<\/p>\n
\n
\u4e3a\u9a8c\u8bc1CoTR\u7684\u6709\u6548\u6027,\u7814\u7a76\u4eba\u5458\u5728\u591a\u9879\u4efb\u52a1\u4e2d\u5c06\u5176\u4e0e\u6807\u51c6\u63d0\u793a\u65b9\u6cd5\u8fdb\u884c\u4e86\u5bf9\u6bd4,\u5305\u62ec\u60c5\u611f\u5206\u6790\u3001\u4ec7\u6068\u8a00\u8bba\u68c0\u6d4b\u3001\u65b0\u95fb\u5206\u7c7b\u548c\u65b0\u95fb\u6807\u9898\u751f\u6210\u7b49\u3002<\/strong><\/span>\u6d4b\u8bd5\u4f7f\u7528\u4e86\u591a\u79cd\u6a21\u578b,\u5982GPT-4o\u3001GPT-4o Mini\u3001Llama 3.1 405B\u548cGemma-9B\u3002<\/p>\n\n
To validate CoTR\u2019s effectiveness, they compared it against standard prompting methods across various tasks \u2014 including\u00a0sentiment analysis, hate speech detection, news categorization, and news headline generation<\/span><\/strong>\u00a0\u2014 using various models such as GPT-4o, GPT-4o Mini,\u00a0Llama 3.1 405B, and Gemma-9B.<\/p>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n\n\n\n03<\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n
\n\n\n\n\u7814\u7a76\u7ed3\u679c\u8868\u660e,\u5c06\u9a6c\u62c9\u5730\u8bed\u8f93\u5165\u53e5\u5b50\u5148\u7ffb\u8bd1\u6210\u82f1\u8bed,\u7136\u540e\u4f7f\u7528\u5355\u4e00\u63d0\u793a\u6267\u884c\u4efb\u52a1,\u6bd4\u76f4\u63a5\u7528\u6807\u51c6\u63d0\u793a\u5904\u7406\u9a6c\u62c9\u5730\u8bed\u6587\u672c\u7684\u6548\u679c\u66f4\u597d\u3002<\/span><\/strong><\/span>CoTR\u5728\u5404\u79cd\u6a21\u578b\u548c\u6570\u636e\u96c6\u4e2d\u7684\u8868\u73b0\u59cb\u7ec8\u4f18\u4e8e\u6807\u51c6\u63d0\u793a\u7b56\u7565\u3002<\/p>\n\n
They found that translating the Marathi input sentence into English and then performing the task using a single prompt yielded superior results compared to directly processing Marathi text with a standard prompt.<\/span><\/strong><\/span>\u00a0CoTR consistently outperformed standard prompting strategies across a variety of models and datasets.<\/p>\n\n
\u7814\u7a76\u4eba\u5458\u8868\u793a:“\u8fd9\u4e9b\u7ed3\u679c\u51f8\u663e\u4e86\u57fa\u4e8e\u7ffb\u8bd1\u7684\u63d0\u793a\u7b56\u7565\u5728\u663e\u8457\u63d0\u5347\u4f4e\u8d44\u6e90\u8bed\u8a00\u591a\u8bed\u8a00LLM\u6027\u80fd\u65b9\u9762\u7684\u6f5c\u529b\u3002”<\/span><\/strong><\/p>\n\n
\u201cThe results underscore the potential of translation-based prompting strategies to significantly improve multilingual LLM performance in low-resource languages,\u201d\u00a0<\/strong><\/span>the researchers said.<\/p>\n\n
\u4ed6\u4eec\u8fd8\u6ce8\u610f\u5230,CoTR\u5728\u8f83\u5c0f\u89c4\u6a21\u6a21\u578b(\u5982Llama3-8B)\u4e0a\u5c55\u73b0\u51fa\u6700\u663e\u8457\u7684\u6027\u80fd\u63d0\u5347\u3002<\/p>\n
\n
They also noted that the most significant performance gains using CoTR were observed with smaller models, such as Llama3-8B.<\/p>\n
\n
\u7814\u7a76\u4eba\u5458\u5f3a\u8c03,\u4ed6\u4eec\u7684\u5de5\u4f5c”\u901a\u8fc7\u5c55\u793a\u57fa\u4e8e\u7ffb\u8bd1\u7684\u63d0\u793a\u7b56\u7565(\u5c24\u5176\u662f\u5355\u4e00\u63d0\u793a)\u5728\u63d0\u5347\u4f4e\u8d44\u6e90\u8bed\u8a00NLP\u6027\u80fd\u65b9\u9762\u7684\u6f5c\u529b,\u4e3a\u591a\u8bed\u8a00NLP\u9886\u57df\u505a\u51fa\u4e86\u91cd\u8981\u8d21\u732e\u3002”<\/span><\/p>\n\n
The researchers highlighted that their work \u201csignificantly contributes to multilingual NLP by demonstrating the potential of translation-based prompting strategies, particularly with a single prompt, to enhance NLP performance in low-resource languages.\u201d<\/span><\/p>\n\n
\u5c55\u671b\u672a\u6765,\u7814\u7a76\u56e2\u961f\u8ba1\u5212\u5c06CoTR\u4e0e\u601d\u7ef4\u94fe\u63d0\u793a(Chain-of-Thought prompting)\u76f8\u7ed3\u5408,\u4ee5\u8fdb\u4e00\u6b65\u63d0\u9ad8\u4f4e\u8d44\u6e90\u8bed\u8a00\u7684NLP\u51c6\u786e\u6027\u3002\u4ed6\u4eec\u8868\u793a:”\u8fd9\u4e9b\u7b56\u7565\u7684\u7ed3\u5408\u6709\u671b\u521b\u5efa\u4e00\u4e2a\u5f3a\u5927\u7684\u6846\u67b6,\u63d0\u5347\u9a6c\u62c9\u5730\u8bedNLP\u4efb\u52a1\u4e2d\u7684\u6a21\u578b\u6027\u80fd\u548c\u53ef\u9760\u6027\u3002<\/span><\/strong><\/p>\n\u00a0<\/span><\/strong><\/p>\nLooking ahead, they plan to combine CoTR with Chain-of-Thought prompting to further improve NLP accuracy for low-resource languages.\u00a0\u201cTogether, these strategies should create a robust framework that improves model performance and reliability in Marathi NLP tasks,\u201d they said.<\/span><\/strong><\/p>\n<\/section>\n<\/section>\n<\/section>\n\n\n<\/section>\n<\/section>\n\n\n\nEND<\/p>\n<\/section>\n<\/section>\n<\/section>\n\n\n\n\n\n\n\n\n\u7279\u522b\u8bf4\u660e\uff1a\u672c\u6587\u5185\u5bb9\u9009\u81ea<\/span>Slator<\/span>\u5b98\u7f51\uff0c\u4ec5\u4f9b\u5b66\u4e60\u4ea4\u6d41\u4f7f\u7528\uff0c\u5982\u6709\u4fb5\u6743\u8bf7\u540e\u53f0\u8054\u7cfb\u5c0f\u7f16\u5220\u9664<\/span>\u3002<\/span><\/p>\n<\/section>\n<\/blockquote>\n<\/section>\n\n
– END –<\/span><\/strong><\/span><\/p>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n<\/section>\n","protected":false},"excerpt":{"rendered":"\u4f4e\u8d44\u6e90\u8bed\u8a00\u7528\u6237\u9700\u6c42\u88ab\u5ffd\u7565\uff1f\u65b0\u6280\u672f\u52a9\u529b\u7f29\u5c0f\u5dee\u8ddd \u539f\u521b\u00a0\u5f90\u5b50\u5170\u00a0\u56fd\u9645\u7ffb\u8bd1\u52a8\u6001 2024\u5e7410\u670827\u65e5 […]<\/p>\n","protected":false},"author":1,"featured_media":35542,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[391],"tags":[],"class_list":["post-35517","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-391"],"_links":{"self":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/35517","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=35517"}],"version-history":[{"count":1,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/35517\/revisions"}],"predecessor-version":[{"id":35545,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/35517\/revisions\/35545"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/media\/35542"}],"wp:attachment":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=35517"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=35517"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=35517"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}