{"id":291164,"date":"2024-11-06T13:47:41","date_gmt":"2024-11-06T12:47:41","guid":{"rendered":"https:\/\/sftarticles.wpenginepowered.com\/es\/?p=343483"},"modified":"2025-07-01T15:46:42","modified_gmt":"2025-07-01T22:46:42","slug":"amd-fully-enters-ai-with-its-first-llm","status":"publish","type":"post","link":"https:\/\/cms-articles.softonic.io\/en\/amd-fully-enters-ai-with-its-first-llm\/","title":{"rendered":"AMD fully enters AI with its first LLM"},"content":{"rendered":"\n<p>AMD has <a href=\"https:\/\/www.amd.com\/en\/developer\/resources\/technical-articles\/introducing-the-first-amd-1b-language-model.html\" target=\"_blank\" rel=\"noopener nofollow\" title=\"\">introduced<\/a> its first series of large language models (LLM) with 1 billion parameters and open source, called <a href=\"https:\/\/huggingface.co\/amd\/AMD-OLMo\" target=\"_blank\" rel=\"noopener nofollow\" title=\"\">AMD OLMo<\/a>, aimed at various applications and pre-trained on the company&#8217;s Instinct MI250 GPUs.<\/p>\n\n\n<div class=\"sc-card-starred-link\">\r\n  <div class=\"sc-card-starred-link__body\">\r\n    <div class=\"sc-card-starred-link__row clearfix\">\r\n      <div class=\"sc-card-starred-link__col-logo\">\r\n        <img decoding=\"async\" class=\"sc-card-starred-link__img\" src=\"https:\/\/articles-img.sftcdn.net\/sft\/articles\/auto-mapping-folder\/sites\/3\/2024\/09\/newsletter.png?GoogleAccessId=wp-stateless%40kubertonic.iam.gserviceaccount.com&Expires=1778885514&Signature=MO6EijRxsc5h7nqY%2FHcwrw1Uv%2FU2xNAbYmV0%2BQOAONWDu7O8UeAAP8pLNSD8Kl6AWky2bsHhq2FKF6tJiZXR12en%2F0cHSTxfPsWbFDqKYxLvT%2Fg2ND7u8l3w3gI3iw43r5E%2FDwaIawEgRGu9wuo7qL8mwhVt%2BKURO4Cmx0SAkil7exV2o6T5%2F9gaLY8eHwMjRV%2Fi3DhXRJai7YSuQPXVL%2Bol8haLGnyNZt7YPrlH3klg2KLZPVyYIqkqvn0M0x3chjUFFfjVmI%2F1eLtECLF8yCIjqx3FSEm2elOF1ihKJwZhIguQGyg0g7rrfo4PvoyFTD7oTRMEZXVOm9O7hGxNwQ%3D%3D\" width=\"100px\" height=\"100px\">\r\n      <\/div>\r\n      <div class=\"sc-card-starred-link__col-title\">\r\n        <p class=\"sc-card-starred-link__title\">Subscribe to the Softonic newsletter and get the latest in tech, gaming, entertainment and deals right in your inbox.<\/p>\r\n        <a class=\"sc-card-starred-link__button\" href=\"https:\/\/softonic.beehiiv.com\/subscribe\" target=\"_blank\" rel=\"noopener noreferrer sponsored\">Subscribe (it's FREE) \u25ba<\/a>\r\n      <\/div>\r\n    <\/div>\r\n    <a class=\"sc-card-starred-link__link\" href=\"https:\/\/softonic.beehiiv.com\/subscribe\" target=\"_blank\" rel=\"noopener noreferrer sponsored\"><\/a>\r\n  <\/div>\r\n<\/div>\n\n\n\n<p><strong>AMD&#8217;s open-source LLMs aim to improve the company&#8217;s position in the AI industry<\/strong> and enable its customers (and everyone) to deploy these open-source models with AMD hardware.<\/p>\n\n\n\n<p>By making the data, weights, training recipes, and code public, <strong>AMD aims to enable developers not only to reproduce the models but also to build upon them to continue innovating<\/strong>.<\/p>\n\n\n\n<p>Beyond use in data centers, AMD has enabled the local deployment of OLMo models on AMD Ryzen AI PCs equipped with neural processing units (NPUs), allowing developers to leverage AI models on personal devices.<\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-rich is-provider-twitter wp-block-embed-twitter\"><div class=\"wp-block-embed__wrapper\">\n<blockquote class=\"twitter-tweet\" data-width=\"550\" data-dnt=\"true\"><p lang=\"en\" dir=\"ltr\">Smol models ftw! <a href=\"https:\/\/twitter.com\/AMD?ref_src=twsrc%5Etfw\">@AMD<\/a> released AMD OLMo 1B &#8211; beats OpenELM, tiny llama on MT Bench, Alpaca Eval &#8211; Apache 2.0 licensed ?<br><br>&gt; Trained with 1.3 trillion (dolma 1.7) tokens on 16 nodes, each with 4 MI250 GPUs<br><br>&gt; Three checkpoints:<br><br>&#8211; AMD OLMo 1B: Pre-trained model<br>&#8211; AMD OLMo 1B SFT:\u2026 <a href=\"https:\/\/t.co\/ae4pLzAKoA\">pic.twitter.com\/ae4pLzAKoA<\/a><\/p>&mdash; Vaibhav (VB) Srivastav (@reach_vb) <a href=\"https:\/\/twitter.com\/reach_vb\/status\/1852764343827648824?ref_src=twsrc%5Etfw\">November 2, 2024<\/a><\/blockquote><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script>\n<\/div><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">Everything we know about AMD&#8217;s LLM<\/h2>\n\n\n\n<p>The AMD OLMo models were trained on a broad dataset of 1.3 trillion tokens across 16 nodes, each with four AMD Instinct MI250 GPUs (64 processors in total). The AMD OLMo model line was trained in three steps.<\/p>\n\n\n\n<p>In AMD&#8217;s own tests, <strong>AMD&#8217;s OLMo models showed impressive performance against similarly sized open-source models,<\/strong> such as TinyLlama-1.1B, MobiLlama-1B, and OpenELM-1_1B in <a href=\"https:\/\/www.tomshardware.com\/tag\/benchmark\" target=\"_blank\" rel=\"noopener\" title=\"\">standard benchmark tests<\/a> for general reasoning capabilities and multitasking comprehension.<\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-rich is-provider-twitter wp-block-embed-twitter\"><div class=\"wp-block-embed__wrapper\">\n<blockquote class=\"twitter-tweet\" data-width=\"550\" data-dnt=\"true\"><p lang=\"en\" dir=\"ltr\">Congratulations to <a href=\"https:\/\/twitter.com\/AMD?ref_src=twsrc%5Etfw\">@AMD<\/a> on the launch of AMD OLMo 1B ? \u201cAMD OLMo 1B is based on the model architecture and training set up of fully open source 1 billion version of OLMo.\u201d Check out more details here: <a href=\"https:\/\/t.co\/Uw0cd5Yn5e\">https:\/\/t.co\/Uw0cd5Yn5e<\/a><\/p>&mdash; Ai2 (@allen_ai) <a href=\"https:\/\/twitter.com\/allen_ai\/status\/1853441318917472361?ref_src=twsrc%5Etfw\">November 4, 2024<\/a><\/blockquote><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script>\n<\/div><\/figure>\n\n\n\n<p>The two-phase SFT model experienced significant improvements in accuracy, with a 5.09% increase in MMLU scores and a 15.32% increase in GSM8k, demonstrating the impact of AMD&#8217;s training approach.<\/p>\n\n\n\n<p><strong>The final model AMD OLMo 1B SFT DPO outperformed other open-source chat models<\/strong> by at least 2.60% on average in benchmark tests.<\/p>\n\n\n<div class=\"sc-card-program\">\r\n  <div class=\"sc-card-program__body\">\r\n    <div class=\"sc-card-program__row clearfix\">\r\n      <div class=\"sc-card-program__col-logo\">\r\n        <img decoding=\"async\" class=\"sc-card-program__img\" alt=\"ChatGPT\" src=\"https:\/\/images.sftcdn.net\/images\/t_app-icon-s\/p\/b330d2b7-464c-4693-b81d-2c97b1edf062\/857405465\/chatgpt-logo\" width=\"100px\" height=\"100px\">\r\n      <\/div>\r\n      <div class=\"sc-card-program__col-title\">\r\n        <span class=\"sc-card-program__title\">ChatGPT<\/span>\r\n        <a class=\"sc-card-program__button sc-card-program-internal\" href=\"https:\/\/chatgpt.en.softonic.com\/iphone\" target=\"_self\" rel=\"noopener noreferrer\">DOWNLOAD<\/a>\r\n      <\/div>\r\n      <div class=\"sc-card-program__col-rating\">\r\n        <svg class=\"rating-score__content\" xmlns=\"http:\/\/www.w3.org\/2000\/svg\" version=\"1.1\" x=\"0\" y=\"0\" viewbox=\"0 0 50 50\" enable-background=\"new 0 0 50 50\" xml:space=\"preserve\"><path class=\"rating-score__background rating-score--good\" fill=\"none\" stroke-width=\"6\" stroke-miterlimit=\"10\" d=\"M40 40c8.3-8.3 8.3-21.7 0-30s-21.7-8.3-30 0 -8.3 21.7 0 30\"><\/path><path class=\"rating-score__value rating-score__value--0\" fill=\"none\" stroke-width=\"6\" stroke-dashoffset=\"0\" stroke-miterlimit=\"10\" d=\"M40 40c8.3-8.3 8.3-21.7 0-30s-21.7-8.3-30 0 -8.3 21.7 0 30\"><\/path><text class=\"rating-score__number\" content=\"\" text-anchor=\"middle\" transform=\"matrix(1 0 0 1 25 31.0837)\" data-auto=\"app-user-score\"><\/text><\/svg>\r\n      <\/div>\r\n    <\/div>\r\n    <div class=\"sc-card-program__row\">\r\n      <span class=\"sc-card-program__description\"><\/span>\r\n    <\/div>\r\n    <div class=\"sc-card-program__row\">\r\n      <img decoding=\"async\" class=\"sc-card-program__bigpic\" src=\"\" onerror=\"this.style.display='none'\">\r\n    <\/div>\r\n    <a class=\"sc-card-program__link track-link sc-card-program-internal\" href=\"https:\/\/chatgpt.en.softonic.com\/iphone\" target=\"_self\" rel=\"noopener noreferrer\"><\/a>\r\n  <\/div>\r\n<\/div>\n\n\n\n<p>Additionally, AMD tested responsible AI tests, such as ToxiGen (which measures toxic language, where a lower score is better), crows_pairs (which evaluates bias), and TruthfulQA-mc2 (which assesses truthfulness in responses). <strong>And it was found that AMD&#8217;s OLMo models were on par with similar models in handling ethical and responsible AI tasks.<\/strong><\/p>\n","protected":false},"excerpt":{"rendered":"<p>AMD has introduced its first series of large language models (LLM) with 1 billion parameters and open source, called AMD OLMo, aimed at various applications and pre-trained on the company&#8217;s Instinct MI250 GPUs. AMD&#8217;s open-source LLMs aim to improve the company&#8217;s position in the AI industry and enable its customers (and everyone) to deploy these open-source models with AMD hardware. By making the data, weights, training recipes, and code public, AMD intends to enable developers not only [&hellip;]<\/p>\n","protected":false},"author":9265,"featured_media":291170,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":"","wpcf-pageviews":2},"categories":[1015],"tags":[3885],"usertag":[],"vertical":[],"content-category":[],"class_list":["post-291164","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news","tag-inteligencia-artificial"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/posts\/291164","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/users\/9265"}],"replies":[{"embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/comments?post=291164"}],"version-history":[{"count":1,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/posts\/291164\/revisions"}],"predecessor-version":[{"id":310923,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/posts\/291164\/revisions\/310923"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/media\/291170"}],"wp:attachment":[{"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/media?parent=291164"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/categories?post=291164"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/tags?post=291164"},{"taxonomy":"usertag","embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/usertag?post=291164"},{"taxonomy":"vertical","embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/vertical?post=291164"},{"taxonomy":"content-category","embeddable":true,"href":"https:\/\/cms-articles.softonic.io\/en\/wp-json\/wp\/v2\/content-category?post=291164"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}