{"id":169794,"date":"2023-08-16T09:23:31","date_gmt":"2023-08-16T14:23:31","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2023\/08\/why-downsizing-large-language-models-is-the-future-of-generative-ai"},"modified":"2023-08-16T09:23:31","modified_gmt":"2023-08-16T14:23:31","slug":"why-downsizing-large-language-models-is-the-future-of-generative-ai","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2023\/08\/why-downsizing-large-language-models-is-the-future-of-generative-ai","title":{"rendered":"Why downsizing large language models is the future of generative AI"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/why-downsizing-large-language-models-is-the-future-of-generative-ai2.jpg\"><\/a><\/p>\n<p>Smaller language models can be based on a billion parameters or less\u2014still pretty large, but much smaller than foundational LLMs like ChatGPT and Bard. They are pre-trained to understand vocabulary and human speech, so the incremental cost to customize them using corporate and industry-specific data is vastly lower. There are several options for these pre-trained LLMs that can be customized internally, including <a href=\"https:\/\/urldefense.proofpoint.com\/v2\/url?u=https-3A__mailtrack.io_trace_link_985d1c46c34a9402bf8ce468f78b9cd5638bc91e-3Furl-3Dhttps-253A-252F-252Fwww.ai21.com-252Fblog-252Fannouncing-2Dai21-2Dstudio-2Dand-2Djurassic-2D1-26userId-3D1276196-26signature-3D4720be17643ef1a4&d=DwMFaQ&c=euGZstcaTDllvimEN8b7jXrwqOf-v5A_CdpgnVfiiMM&r=p8bAs6sMIJAg-G2lyHfE7aKsn50wUzJqF0tQ6l8h4ng&m=VQZT082E1GUA1QJTqepkth2rLFcUOUtXql4k1Tr5iE5H2EtA63IHIKRYL8xcyjO9&s=gU5Qdkw96RMWojzBiuE7Z6iIEIFf7Tc_qIiX1Ia44Yk\">AI21<\/a> and <a href=\"https:\/\/urldefense.proofpoint.com\/v2\/url?u=https-3A__mailtrack.io_trace_link_fae9a11bbdab6f92ecfe4f447622db671e5436f1-3Furl-3Dhttps-253A-252F-252Freka.ai-252F-26userId-3D1276196-26signature-3D48f955d197aac29a&d=DwMFaQ&c=euGZstcaTDllvimEN8b7jXrwqOf-v5A_CdpgnVfiiMM&r=p8bAs6sMIJAg-G2lyHfE7aKsn50wUzJqF0tQ6l8h4ng&m=VQZT082E1GUA1QJTqepkth2rLFcUOUtXql4k1Tr5iE5H2EtA63IHIKRYL8xcyjO9&s=Ib409YfE3x_R3kttXXeNv2v2mWqB0NXUfVSPJvUd7rY\">Reka<\/a>, as well as open source LLMs like <a href=\"https:\/\/urldefense.proofpoint.com\/v2\/url?u=https-3A__mailtrack.io_trace_link_3dea11a641e5113cd37572e209b2bc5caee943cd-3Furl-3Dhttps-253A-252F-252Fcrfm.stanford.edu-252F2023-252F03-252F13-252Falpaca.html-26userId-3D1276196-26signature-3D7b0a77ca1dcabc5c&d=DwMFaQ&c=euGZstcaTDllvimEN8b7jXrwqOf-v5A_CdpgnVfiiMM&r=p8bAs6sMIJAg-G2lyHfE7aKsn50wUzJqF0tQ6l8h4ng&m=VQZT082E1GUA1QJTqepkth2rLFcUOUtXql4k1Tr5iE5H2EtA63IHIKRYL8xcyjO9&s=X9_hoDDHMjiJPoC8zBkP44VBUK4oEWBXIAiE0qsB3SU\">Alpaca<\/a> and <a href=\"https:\/\/urldefense.proofpoint.com\/v2\/url?u=https-3A__mailtrack.io_trace_link_67c481b193743760aa2bc729b5851ff5d3b074b9-3Furl-3Dhttps-253A-252F-252Flmsys.org-252Fblog-252F2023-2D03-2D30-2Dvicuna-252F-26userId-3D1276196-26signature-3D0e83962fcba2586f&d=DwMFaQ&c=euGZstcaTDllvimEN8b7jXrwqOf-v5A_CdpgnVfiiMM&r=p8bAs6sMIJAg-G2lyHfE7aKsn50wUzJqF0tQ6l8h4ng&m=VQZT082E1GUA1QJTqepkth2rLFcUOUtXql4k1Tr5iE5H2EtA63IHIKRYL8xcyjO9&s=lb69Z-DxfgleGNxhga5nZ8xMH4Pv0Rgs7CeeOyKxI94\">Vicuna<\/a>.<\/p>\n<p>Smaller language models aren\u2019t just more cost-efficient, they\u2019re often far more accurate, because instead of training them on all publicly available data\u2014the good and the bad\u2014they are trained and optimized on carefully vetted data that addresses the exact use cases a business cares about.<\/p>\n<p>That doesn\u2019t mean they\u2019re limited to internal corporate data. Smaller language models can incorporate third-party data about the economy, commodities pricing, the weather, or whatever data sets are needed, and combine them with their proprietary data sets. These data sources are widely available from data service providers who ensure the information is current, accurate, and clean.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Smaller language models can be based on a billion parameters or less\u2014still pretty large, but much smaller than foundational LLMs like ChatGPT and Bard. They are pre-trained to understand vocabulary and human speech, so the incremental cost to customize them using corporate and industry-specific data is vastly lower. There are several options for these pre-trained [\u2026]<\/p>\n","protected":false},"author":556,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[43,39,6],"tags":[],"class_list":["post-169794","post","type-post","status-publish","format-standard","hentry","category-business","category-economics","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/169794","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/556"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=169794"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/169794\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=169794"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=169794"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=169794"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}