Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the betterdocs domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /data/user/htdocs/wp-includes/functions.php on line 6114

Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the jnews-view-counter domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /data/user/htdocs/wp-includes/functions.php on line 6114

Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the wp-statistics domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /data/user/htdocs/wp-includes/functions.php on line 6114

Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the wpdiscuz domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /data/user/htdocs/wp-includes/functions.php on line 6114

Notice: 函数 _load_textdomain_just_in_time 的调用方法不正确jnews 域的翻译加载触发过早。这通常表示插件或主题中的某些代码运行过早。翻译应在 init 操作或之后加载。 请查阅调试 WordPress来获取更多信息。 (这个消息是在 6.7.0 版本添加的。) in /data/user/htdocs/wp-includes/functions.php on line 6114

Notice: 函数 _load_textdomain_just_in_time 的调用方法不正确jnews-like 域的翻译加载触发过早。这通常表示插件或主题中的某些代码运行过早。翻译应在 init 操作或之后加载。 请查阅调试 WordPress来获取更多信息。 (这个消息是在 6.7.0 版本添加的。) in /data/user/htdocs/wp-includes/functions.php on line 6114

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893

Warning: Cannot modify header information - headers already sent by (output started at /data/user/htdocs/wp-includes/functions.php:6114) in /data/user/htdocs/wp-includes/rest-api/class-wp-rest-server.php on line 1893
{"id":33048,"date":"2024-08-29T17:34:49","date_gmt":"2024-08-29T09:34:49","guid":{"rendered":"https:\/\/linguaresources.com\/?p=33048"},"modified":"2024-08-29T17:34:49","modified_gmt":"2024-08-29T09:34:49","slug":"how-well-does-llama-3-1-perform-for-text-and-speech-translation","status":"publish","type":"post","link":"https:\/\/linguaresources.com\/?p=33048","title":{"rendered":"How Well Does Llama 3.1 Perform for Text and Speech Translation?"},"content":{"rendered":"

<\/p>\n

\n

Meta\u2019s research team introduced<\/a> Llama 3.1 on July 23, 2023, calling it \u201cthe world\u2019s largest and most capable openly available foundation model.\u201d<\/p>\n

Llama 3.1 is available in various parameter sizes \u2014 8B, 70B, and 405B \u2014 providing flexibility for deployment based on computational resources and specific application needs. On April 18, 2024, Meta announced<\/a> the Llama 3 family of large language models<\/a>, which initially included only the 8B and 70B sizes. This latest release introduced the 405B model along with upgraded versions of the 8B and 70B models.<\/p>\n

Llama 3.1 models represent a significant advancement over their predecessor, Llama 2, being pre-trained on an extensive corpus of 15 trillion multilingual tokens, a substantial increase from Llama 2\u2019s 1.8 trillion tokens. With a context window of up to 128k tokens \u2014 previously limited to 8k tokes \u2014 they offer notable improvements in multilinguality, coding, reasoning, and tool usage.<\/p>\n

if (!window.AdButler){(function(){var s = document.createElement(“script”); s.async = true; s.type = “text\/javascript”;s.src = ‘https:\/\/servedbyadbutler.com\/app.js’;var n = document.getElementsByTagName(“script”)[0]; n.parentNode.insertBefore(s, n);}());}<\/p>\n
<\/div>\n

var AdButler = AdButler || {}; AdButler.ads = AdButler.ads || [];
\nvar abkw = window.abkw || ”;
\nvar plc336548 = window.plc336548 || 0;
\n(function(){
\nvar divs = document.querySelectorAll(“.plc336548:not([id])”);
\nvar div = divs[divs.length-1];
\ndiv.id = “placement_336548_”+plc336548;
\nAdButler.ads.push({handler: function(opt){ AdButler.register(167169, 336548, [728,90], ‘placement_336548_’+opt.place, opt); }, opt: { place: plc336548++, keywords: abkw, domain: ‘servedbyadbutler.com’, click:’CLICK_MACRO_PLACEHOLDER’ }});
\n})(); <\/p><\/div>\n

Llama 3.1 maintains a similar architecture to Llama and Llama 2 but achieves performance improvements through enhanced data quality, diversity, and increased training scale.\u00a0<\/p>\n

Meta\u2019s research team tested Llama 3.1 on over 150 benchmark datasets covering a wide range of languages. They found that their \u201cflagship model\u201d with 405B parameters is competitive with leading models<\/a> across various tasks and is close to matching the state-of-the-art performance. The smaller models are also \u201cbest-in-class,\u201d outperforming alternative models with comparable numbers of parameters.\u00a0<\/p>\n

SOTA Capabilities in Multilingual Translation<\/h3>\n

In multilingual tasks, the small Llama 3.1 8B model surpassed Gemma 2 9B and Mistral 7B, while Llama 3.1 70B outperformed Mixtral 8Xx22B and GPT 3.5 Turbo. Llama 3.1 405B is on par with Claude 3.5 Sonnet and outperformed GPT-4<\/a> and GPT 4o<\/a>.<\/p>\n

Meta\u2019s research team emphasized that Llama 3.1 405B is \u201cthe first openly available model that rivals the top AI models when it comes to state-of-the-art capabilities in [\u2026] multilingual translation,\u201d among other tasks.<\/p>\n

They expressed optimism about the potential for creating innovative applications leveraging the model\u2019s multilingual capabilities and extended context length, stating, \u201cwe can\u2019t wait to see what the community does with this work.\u201d.<\/p>\n

if (!window.AdButler){(function(){var s = document.createElement(“script”); s.async = true; s.type = “text\/javascript”;s.src = ‘https:\/\/servedbyadbutler.com\/app.js’;var n = document.getElementsByTagName(“script”)[0]; n.parentNode.insertBefore(s, n);}());}<\/p>\n
<\/div>\n

var AdButler = AdButler || {}; AdButler.ads = AdButler.ads || [];
\nvar abkw = window.abkw || ”;
\nvar plc332285 = window.plc332285 || 0;
\n(function(){
\nvar divs = document.querySelectorAll(“.plc332285:not([id])”);
\nvar div = divs[divs.length-1];
\ndiv.id = “placement_332285_”+plc332285;
\nAdButler.ads.push({handler: function(opt){ AdButler.register(167169, 332285, [728,90], ‘placement_332285_’+opt.place, opt); }, opt: { place: plc332285++, keywords: abkw, domain: ‘servedbyadbutler.com’, click:’CLICK_MACRO_PLACEHOLDER’ }});
\n})(); <\/p><\/div>\n

Strong Performance on Speech Translation<\/h3>\n

In addition to language processing, the development of Llama 3.1 included multimodal extensions that enable image recognition, video recognition, and speech understanding capabilities.<\/p>\n

Although these multimodal extensions are still under development, initial results indicate competitive performance in image, video, and speech tasks.<\/p>\n

Meta\u2019s research team specifically evaluated Llama 3.1 on automatic speech recognition<\/a> (ASR) and speech translation<\/a>. In ASR<\/a>, they compared its performance against Whisper<\/a>, SeamlessM4T, and Gemini. Llama 3.1 outperformed Whisper and SeamlessM4T across all benchmarks and performed similarly to Gemini, demonstrating \u201cstrong performance on speech recognition<\/a> tasks.\u201d\u00a0<\/p>\n

\n
\n
\n
\n
\"10<\/div>\n

Slator Pro Guide: Translation AI<\/h2>\n
The Slator Pro Guide presents 10 new and impactful ways that LLMs can be used to enhance translation workflows.<\/span><\/div>\n
$<\/span>290<\/span><\/span> BUY NOW<\/a>Included in our Pro and Enterprise plan.
Subscribe now!<\/a><\/span><\/div>\n<\/div>\n<\/article>\n<\/div>\n<\/div>\n

In speech translation tasks, where the model was asked to translate non-English speech into English text, Llama 3.1 again outperformed Whisper and SeamlesM4T. \u201cThe performance of our models in speech translation highlights the advantages of multimodal foundation models for tasks such as speech translation,\u201d Meta\u2019s team said.<\/p>\n

They also shared details of the development process to help the research community understand the key factors of multimodal foundation model development and encourage informed discussions about the future of these models. \u201cWe hope sharing our results early will accelerate research in this direction,\u201d they said.<\/p>\n

Early Use Cases<\/h3>\n

Meta\u2019s launch of Llama 3.1 has created a buzz in the AI community. Since the release, many people have taken to X and LinkedIn to call it a \u201cgame-changer<\/a>\u201d or \u201cGPT-4 killer<\/a>,\u201d recognizing this moment as \u201cthe biggest moment for open-source AI<\/a>.\u201d Additionally, they have talked about a \u201cseismic shift in business transformation<\/a>,\u201d explaining that this is going to \u201crevolutionize how companies work.\u201d\u00a0<\/p>\n

Posts are filled with examples showing the many different ways Llama 3.1 can be used<\/a>, building from phone assistants<\/a> to document assistants<\/a> and code assistants<\/a>.<\/p>\n

\n
\n
\n
\n

Groq + LLaMa 3.1-8b is just too much fun.<\/p>\n

People are sharing instant responses from voice notes.
I tried it myself & it’s wild:
pic.twitter.com\/yWimJhPZuC<\/a><\/p>\n

\u2014 Ruben Hassid (@RubenHssd) July 25, 2024<\/a><\/p><\/blockquote>\n<\/div>\n<\/div>\n<\/figure>\n

Publicly Available<\/h3>\n

Meta has released all Llama 3.1 models under an updated community license, promoting further innovation and responsible development towards artificial general intelligence (AGI).<\/p>\n

\u201cWe hope that the open release of a flagship model will spur a wave of innovation in the research community, and accelerate a responsible path towards the development of artificial general intelligence\u201d they said. Additionally, they believe that the release of Llama 3.1 will encourage the industry to adopt open and responsible practices in AGI development.<\/p>\n

The Meta research team acknowledges that there is still much to explore, including more device-friendly sizes, additional modalities, and further investment in the agent platform layer.<\/p>\n

The models are available for download on llama.meta.com<\/a> and Hugging Face<\/a> and ready for immediate development within a broad ecosystem of partner platforms, including AWS, NVIDIA, Databricks, Groq, Dell, Azure, Google Cloud, and Snowflake.<\/p>\n

Ahmad Al-Dahle, who leads Meta\u2019s generative AI efforts, wrote in a post on X<\/a>, \u201cWith Llama 3.1 in NVIDIA AI Foundry we\u2019ll see enterprises to easily create custom AI services with the world\u2019s best open source AI models.\u201d<\/p>\n<\/div>\n

\u539f\u6587\u94fe\u63a5 <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"

Meta\u2019s research team introduced Llama 3.1 on July 23, 2 […]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[391],"tags":[],"class_list":["post-33048","post","type-post","status-publish","format-standard","hentry","category-391"],"_links":{"self":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/33048","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=33048"}],"version-history":[{"count":0,"href":"https:\/\/linguaresources.com\/index.php?rest_route=\/wp\/v2\/posts\/33048\/revisions"}],"wp:attachment":[{"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=33048"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=33048"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/linguaresources.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=33048"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}