Meta\u2019s research team introduced<\/a> Llama 3.1 on July 23, 2023, calling it \u201cthe world\u2019s largest and most capable openly available foundation model.\u201d<\/p>\n Llama 3.1 is available in various parameter sizes \u2014 8B, 70B, and 405B \u2014 providing flexibility for deployment based on computational resources and specific application needs. On April 18, 2024, Meta announced<\/a> the Llama 3 family of large language models<\/a>, which initially included only the 8B and 70B sizes. This latest release introduced the 405B model along with upgraded versions of the 8B and 70B models.<\/p>\n Llama 3.1 models represent a significant advancement over their predecessor, Llama 2, being pre-trained on an extensive corpus of 15 trillion multilingual tokens, a substantial increase from Llama 2\u2019s 1.8 trillion tokens. With a context window of up to 128k tokens \u2014 previously limited to 8k tokes \u2014 they offer notable improvements in multilinguality, coding, reasoning, and tool usage.<\/p>\n var AdButler = AdButler || {}; AdButler.ads = AdButler.ads || []; Llama 3.1 maintains a similar architecture to Llama and Llama 2 but achieves performance improvements through enhanced data quality, diversity, and increased training scale.\u00a0<\/p>\n Meta\u2019s research team tested Llama 3.1 on over 150 benchmark datasets covering a wide range of languages. They found that their \u201cflagship model\u201d with 405B parameters is competitive with leading models<\/a> across various tasks and is close to matching the state-of-the-art performance. The smaller models are also \u201cbest-in-class,\u201d outperforming alternative models with comparable numbers of parameters.\u00a0<\/p>\n In multilingual tasks, the small Llama 3.1 8B model surpassed Gemma 2 9B and Mistral 7B, while Llama 3.1 70B outperformed Mixtral 8Xx22B and GPT 3.5 Turbo. Llama 3.1 405B is on par with Claude 3.5 Sonnet and outperformed GPT-4<\/a> and GPT 4o<\/a>.<\/p>\n Meta\u2019s research team emphasized that Llama 3.1 405B is \u201cthe first openly available model that rivals the top AI models when it comes to state-of-the-art capabilities in [\u2026] multilingual translation,\u201d among other tasks.<\/p>\n They expressed optimism about the potential for creating innovative applications leveraging the model\u2019s multilingual capabilities and extended context length, stating, \u201cwe can\u2019t wait to see what the community does with this work.\u201d.<\/p>\n var AdButler = AdButler || {}; AdButler.ads = AdButler.ads || []; In addition to language processing, the development of Llama 3.1 included multimodal extensions that enable image recognition, video recognition, and speech understanding capabilities.<\/p>\n Although these multimodal extensions are still under development, initial results indicate competitive performance in image, video, and speech tasks.<\/p>\n Meta\u2019s research team specifically evaluated Llama 3.1 on automatic speech recognition<\/a> (ASR) and speech translation<\/a>. In ASR<\/a>, they compared its performance against Whisper<\/a>, SeamlessM4T, and Gemini. Llama 3.1 outperformed Whisper and SeamlessM4T across all benchmarks and performed similarly to Gemini, demonstrating \u201cstrong performance on speech recognition<\/a> tasks.\u201d\u00a0<\/p>\n
\nvar abkw = window.abkw || ”;
\nvar plc336548 = window.plc336548 || 0;
\n(function(){
\nvar divs = document.querySelectorAll(“.plc336548:not([id])”);
\nvar div = divs[divs.length-1];
\ndiv.id = “placement_336548_”+plc336548;
\nAdButler.ads.push({handler: function(opt){ AdButler.register(167169, 336548, [728,90], ‘placement_336548_’+opt.place, opt); }, opt: { place: plc336548++, keywords: abkw, domain: ‘servedbyadbutler.com’, click:’CLICK_MACRO_PLACEHOLDER’ }});
\n})(); <\/p><\/div>\nSOTA Capabilities in Multilingual Translation<\/h3>\n
\nvar abkw = window.abkw || ”;
\nvar plc332285 = window.plc332285 || 0;
\n(function(){
\nvar divs = document.querySelectorAll(“.plc332285:not([id])”);
\nvar div = divs[divs.length-1];
\ndiv.id = “placement_332285_”+plc332285;
\nAdButler.ads.push({handler: function(opt){ AdButler.register(167169, 332285, [728,90], ‘placement_332285_’+opt.place, opt); }, opt: { place: plc332285++, keywords: abkw, domain: ‘servedbyadbutler.com’, click:’CLICK_MACRO_PLACEHOLDER’ }});
\n})(); <\/p><\/div>\nStrong Performance on Speech Translation<\/h3>\n