{"id":115,"date":"2024-03-24T13:47:31","date_gmt":"2024-03-24T13:47:31","guid":{"rendered":"https:\/\/wsw-int.de\/?p=115"},"modified":"2024-10-25T13:24:15","modified_gmt":"2024-10-25T13:24:15","slug":"techniques-for-using-language-models","status":"publish","type":"post","link":"https:\/\/multai.eu\/de\/techniques-for-using-language-models\/","title":{"rendered":"Techniken zur Nutzung von Sprachmodellen"},"content":{"rendered":"<p>One of the weaknesses of the models currently available on the market is that they have been trained on a publicly accessible data set, which may not necessarily be sufficient to meet certain specific needs.<\/p>\n\n\n\n<p>Take, for example, a company with a large volume of proprietary data, a highly specialized vocabulary or specific data formats. This knowledge will not a priori be integrated into a general-purpose model, as the data is not publicly accessible. What can be done about this? That&#8217;s what I&#8217;d like to talk about today.<\/p>\n\n\n\n<p>There are several techniques for &#8220;enriching&#8221; model knowledge. These include, in order of increasing complexity<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>prompt engineering;<\/li>\n\n\n\n<li>augmented recovery generation (RAG);<\/li>\n\n\n\n<li>model refinement (complete or optimized).<\/li>\n<\/ul>\n\n\n\n<p>These techniques can be found in the middle of the figure below, between the two extremes of simple dialogue and complete training:<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img fetchpriority=\"high\" decoding=\"async\" width=\"1024\" height=\"578\" src=\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\" alt=\"\" class=\"wp-image-116\" srcset=\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp 1024w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218a-300x169.webp 300w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218a-768x434.webp 768w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218a-1536x867.webp 1536w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218a.webp 2046w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Figure 1: Techniques for using language models<\/figcaption><\/figure>\n\n\n\n<p>This is an interesting topic to cover now, as it complements well our <a href=\"https:\/\/wsw-int.de\/local-generative-models\">previous article<\/a> on local models. Indeed, certain techniques such as refinement are mainly aimed at open-source models whose parameters are freely available.<\/p>\n\n\n\n<p>Let&#8217;s take a look at the different techniques.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">1. Prompt Engineering<\/h2>\n\n\n\n<p>Prompt engineering comprises two basic underlying techniques: the first is to provide sufficient contextual information in the prompt. The model will then combine all the information in its possession (from the training and the prompt) to answer the question posed.<\/p>\n\n\n\n<p>The second technique involves encouraging the model to follow structured logical reasoning. Since language models have been trained to comply with the user&#8217;s instructions, they can be asked to follow a step-by-step reasoning process or to provide a series of solved examples to put the model on the right track. While not a panacea, these techniques have already proved effective.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" width=\"1024\" height=\"456\" src=\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218b-1024x456.webp\" alt=\"\" class=\"wp-image-117\" srcset=\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218b-1024x456.webp 1024w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218b-300x133.webp 300w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218b-768x342.webp 768w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218b-1536x683.webp 1536w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218b.webp 2032w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Figure 2: Prompt Engineering<\/figcaption><\/figure>\n\n\n\n<p>A fun method is to promise a financial reward to the model if they answer correctly. Simply add &#8220;If you answer correctly, I&#8217;ll give you ten euros&#8221; at the end of the prompt. Believe it or not, it seems to work!<\/p>\n\n\n\n<p>Prompt engineering can provide sufficient precision in some applications. Its main advantage is its simplicity of implementation, which also requires no additional computing resources.<\/p>\n\n\n\n<p>But this technique comes up against the limited size of context windows in language models. It cannot be applied when the volume of information to be transmitted is too great or too complex to describe in the prompt. There is then a risk that the model will invent the missing information, resulting in incorrect or even nonsensical responses (hallucinations).<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">2. Retrieval Augmented Generation (RAG)<\/h2>\n\n\n\n<p>The idea may have occurred to you on reading the previous point: why not combine prompt engineering with a search engine that indexes additional data? The result would be a prompt &#8220;enriched&#8221; with the most significant additional elements, automatically and transparently for the user\u2026<\/p>\n\n\n\n<p>This is the promise of Retrieval Augmented Generation (RAG). Here&#8217;s how it works:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Additional information is broken down into blocks, each of which is indexed according to its content. These indexes are usually vectors whose position in space depends on the content of the block (semantic indexing). An introduction to this subject can be found here;<\/li>\n\n\n\n<li>All the indexes are stored in a vector database, which also contains the references of the indexed text blocks;<\/li>\n\n\n\n<li>When a user asks a question, the question text is also indexed using the same algorithm. A search of the vector database will identify blocks semantically close to the prompt based on proximity between vectors;<\/li>\n\n\n\n<li>These semantically close blocks of information are concatenated with the original prompt as additional context;<\/li>\n\n\n\n<li>The prompt enriched with the additional context is sent to the language model for response.<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" width=\"1024\" height=\"416\" src=\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218c-1024x416.webp\" alt=\"\" class=\"wp-image-118\" srcset=\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218c-1024x416.webp 1024w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218c-300x122.webp 300w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218c-768x312.webp 768w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218c-1536x624.webp 1536w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218c.webp 2046w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Figure 3: Retrieval Augmented Generation (RAG)<\/figcaption><\/figure>\n\n\n\n<p><\/p>\n\n\n\n<p>This technique has several advantages. Imagine a typical set of proprietary information generated by a company: databases, PDF documents, excel files, news feeds, meeting minutes\u2026.. All this data can be sliced, semantically indexed, and fed into the vector database. And it&#8217;s easy enough to continue enriching the vector database regularly to ensure that the information remains up to date.<\/p>\n\n\n\n<p>The second major advantage is that there&#8217;s no need to modify the language model as such. It all depends, of course, on the efficiency of the semantic indexing mechanism, but after 25 years of Internet search engines, these techniques are mature.<\/p>\n\n\n\n<p>What&#8217;s more, a programming library like Langchain contains all the functionality needed to orchestrate interaction with the vector database and language model. In particular, this library supports a hundred or so file formats to feed the knowledge base.<\/p>\n\n\n\n<p>The price to pay is greater complexity in the IT architecture. Several infrastructure elements have to be integrated and combined. And if a cloud-based vector database like Pinecone is used, the confidentiality risks that go with it must also be taken into account.<\/p>\n\n\n\n<p>Finally, as context enrichment is a one-off, targeted process, this technique is not appropriate if you&#8217;re aiming to specialize a language model in a complex field such as medicine or finance.<\/p>\n\n\n\n<p>In this case, it&#8217;s better to refine the model. That&#8217;s what we&#8217;re going to look at now.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">3. Model Fine-Tuning<\/h2>\n\n\n\n<p>Model fine-tuning is a process during which an already pre-trained model undergoes additional training on a specific data set. This process capitalizes on the knowledge already integrated during initial training, reinforcing the model&#8217;s competence in a specific domain in return for a reasonable investment in computing resources.<\/p>\n\n\n\n<p>This technique is important because the initial training of a model requires enormous resources, making it out of reach for most organizations.<\/p>\n\n\n\n<p>Power requirements can be further reduced by using advanced refinement techniques such as LoRA (Low-Rank Adaptation). Introduced at the beginning of 2023, this method considerably reduces the number of parameters to be trained, at the cost of a slight degradation in quality.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"402\" src=\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218d-1024x402.webp\" alt=\"\" class=\"wp-image-119\" style=\"width:670px;height:auto\" srcset=\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218d-1024x402.webp 1024w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218d-300x118.webp 300w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218d-768x302.webp 768w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218d-1536x603.webp 1536w, https:\/\/multai.eu\/wp-content\/uploads\/2024\/03\/image-20240218d.webp 2048w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Figure 4: Optimized fine-tuning<\/figcaption><\/figure>\n\n\n\n<p>You&#8217;ll need a sufficiently large data set for further training. If you&#8217;re looking to deepen the model&#8217;s expertise in a specific domain, you can start by using any reference material on the subject you have available.<\/p>\n\n\n\n<p>In addition, you can improve the model&#8217;s response type by adding a set of prompts and their responses to the training data. This data set may have been generated manually or via a &#8220;high-end&#8221; language model such as GPT4.<\/p>\n\n\n\n<p>In any case, refinement remains a more complex technique, requiring sufficient AI expertise to build the dataset, implement the model training algorithm, and then evaluate the performance of the modified model. This approach also requires access to significant computing power.<\/p>\n\n\n\n<p>One drawback of this approach compared with Augmented Recovery Generation is that it is much more difficult to introduce new information into the model: you have to go through a training phase again, with all the effort that this entails.<\/p>\n\n\n\n<p>Another constraint of this approach is that, to modify a model, its parameters must be available. In practice, therefore, only open-source models such as Llama2 or Mistral lend themselves to this kind of exercise.<\/p>\n\n\n\n<p>Finally, it&#8217;s worth noting that refined versions of open-source language models like Llama are available on the Internet, for specific fields such as programming. Using such a model can also be a solution\u2026<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">4. Reflections<\/h2>\n\n\n\n<p>The advanced exploitation of language models described above is progressing rapidly. Optimized&#8221; refinement techniques, semantic search algorithms and vector-based RAG databases are all making steady progress.<\/p>\n\n\n\n<p>Techniques such as RAG or refinement are too cumbersome for private use, but are an interesting solution for companies. The availability of open-source models combined with these techniques offers great deployment flexibility to organizations wishing to exploit language models to the best of their ability.<\/p>\n\n\n\n<p>And the ability to run the whole thing &#8220;in-house&#8221; offers an elegant answer to the confidentiality concerns that hold many organizations back.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">5. Notes and references<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li><em>Full Fine-Tuning, PEFT, Prompt Engineering and RAG: Which One is Right for You<\/em> ?, by Najeeb Nawani for Deci.ai, Sept. 20th 2023: <a href=\"https:\/\/deci.ai\/blog\/fine-tuning-peft-prompt-engineering-and-rag-which-one-is-right-for-you\/\">https:\/\/deci.ai\/blog\/fine-tuning-peft-prompt-engineering-and-rag-which-one-is-right-for-you\/<\/a><\/li>\n\n\n\n<li><em>Qu\u2019est-ce que la g\u00e9n\u00e9ration augment\u00e9e de r\u00e9cup\u00e9ration (RAG, retrieval-augmented generation)&nbsp;?<\/em>, by Alan Zichik for Oracle, Sept. 19th 2023<em>: <\/em><a href=\"https:\/\/www.oracle.com\/fr\/artificial-intelligence\/generative-ai\/retrieval-augmented-generation-rag\/\">https:\/\/www.oracle.com\/fr\/artificial-intelligence\/generative-ai\/retrieval-augmented-generation-rag\/<\/a><\/li>\n\n\n\n<li><em>LLM Explained: The LLM Training Landscape<\/em>, by Crystal Liu on Medium, Aug. 7th, 2023: <a href=\"https:\/\/liu-gendary.medium.com\/llm-explained-the-llm-training-landscape-82c803495caa\">https:\/\/liu-gendary.medium.com\/llm-explained-the-llm-training-landscape-82c803495caa<\/a><\/li>\n\n\n\n<li><em>Making LLMs Work For Your Use-Case \u2013 Fine-Tuning vs. RAG vs. Long Context Length vs. Prompting<\/em>, tweet by Bindu Reddy: <a href=\"https:\/\/twitter.com\/bindureddy\/status\/1712257709918318898\">https:\/\/twitter.com\/bindureddy\/status\/1712257709918318898<\/a><\/li>\n\n\n\n<li><em>The Art of Fine-Tuning Large Language Models<\/em>, tweet by Bindu Reddy: <a href=\"https:\/\/twitter.com\/bindureddy\/status\/1699275289493430699\">https:\/\/twitter.com\/bindureddy\/status\/1699275289493430699<\/a><\/li>\n\n\n\n<li><em>Vector Embeddings for Developers: The Basics<\/em>, par Roie Schwaber-Cohen on Pinecone.io, June 30th 2023: <a href=\"https:\/\/www.pinecone.io\/learn\/vector-embeddings-for-developers\/\">https:\/\/www.pinecone.io\/learn\/vector-embeddings-for-developers\/<\/a><\/li>\n<\/ul>\n\n\n\n<p><a href=\"https:\/\/multai.eu\/de\/\">MultAI.eu<\/a> &#8230;<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" id=\"block-558d25e1-1f38-48e1-90f0-ab4fec0a05f5\"\/>\n\n\n\n<p id=\"block-d4bfcc52-6a5e-4883-a1f5-1461e662e3d9\"><mark><mark style=\"background-color:#ffffff\" class=\"has-inline-color\">Translated with <a href=\"https:\/\/www.deepl.com\/translator\">DeepL <\/a>and adapted from our partner Arnaud Stevins&#8217; <a href=\"https:\/\/artificiellementintelligent.wordpress.com\/2024\/03\/18\/des-modeles-de-langage-aux-modeles-multimodaux\/\">blog<\/a> (February. 10th, 2024).<\/mark><\/mark><\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" id=\"block-558d25e1-1f38-48e1-90f0-ab4fec0a05f5\"\/>\n\n\n\n<p>March 24th, 2024<\/p>\n\n\n\n<p><\/p>","protected":false},"excerpt":{"rendered":"<p>One of the weaknesses of the models currently available on the market is that they have been trained on a publicly accessible data set, which may not necessarily be sufficient to meet certain specific needs. Take, for example, a company with a large volume of proprietary data, a highly specialized vocabulary or specific data formats. [&hellip;]<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[25,19,10,26,22,23,24],"class_list":["post-115","post","type-post","status-publish","format-standard","hentry","category-uncategorized","tag-finetuning","tag-genai","tag-llm","tag-prompt","tag-prompting","tag-rag","tag-retrieval-augmented-generation"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v22.5 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Techniques for using language models - MultAI<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/multai.eu\/de\/techniques-for-using-language-models\/\" \/>\n<meta property=\"og:locale\" content=\"de_DE\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Techniques for using language models - MultAI\" \/>\n<meta property=\"og:description\" content=\"One of the weaknesses of the models currently available on the market is that they have been trained on a publicly accessible data set, which may not necessarily be sufficient to meet certain specific needs. Take, for example, a company with a large volume of proprietary data, a highly specialized vocabulary or specific data formats. [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/multai.eu\/de\/techniques-for-using-language-models\/\" \/>\n<meta property=\"og:site_name\" content=\"MultAI\" \/>\n<meta property=\"article:published_time\" content=\"2024-03-24T13:47:31+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-10-25T13:24:15+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\" \/>\n<meta name=\"author\" content=\"hans\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Verfasst von\" \/>\n\t<meta name=\"twitter:data1\" content=\"hans\" \/>\n\t<meta name=\"twitter:label2\" content=\"Gesch\u00e4tzte Lesezeit\" \/>\n\t<meta name=\"twitter:data2\" content=\"7\u00a0Minuten\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/\"},\"author\":{\"name\":\"hans\",\"@id\":\"https:\/\/multai.eu\/#\/schema\/person\/06def8c374b5d6724bec911e9880c292\"},\"headline\":\"Techniques for using language models\",\"datePublished\":\"2024-03-24T13:47:31+00:00\",\"dateModified\":\"2024-10-25T13:24:15+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/\"},\"wordCount\":1429,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/multai.eu\/#organization\"},\"image\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\",\"keywords\":[\"finetuning\",\"GenAI\",\"LLM\",\"prompt\",\"prompting\",\"RAG\",\"retrieval-augmented generation\"],\"articleSection\":[\"Uncategorized\"],\"inLanguage\":\"de\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/multai.eu\/techniques-for-using-language-models\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/\",\"url\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/\",\"name\":\"Techniques for using language models - MultAI\",\"isPartOf\":{\"@id\":\"https:\/\/multai.eu\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\",\"datePublished\":\"2024-03-24T13:47:31+00:00\",\"dateModified\":\"2024-10-25T13:24:15+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#breadcrumb\"},\"inLanguage\":\"de\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/multai.eu\/techniques-for-using-language-models\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"de\",\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage\",\"url\":\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\",\"contentUrl\":\"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/multai.eu\/techniques-for-using-language-models\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/multai.eu\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Techniques for using language models\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/multai.eu\/#website\",\"url\":\"https:\/\/multai.eu\/\",\"name\":\"WSW\",\"description\":\"Generative AI for your business\",\"publisher\":{\"@id\":\"https:\/\/multai.eu\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/multai.eu\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"de\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/multai.eu\/#organization\",\"name\":\"WSW\",\"alternateName\":\"MultAI\",\"url\":\"https:\/\/multai.eu\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"de\",\"@id\":\"https:\/\/multai.eu\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/10\/Logo.png\",\"contentUrl\":\"https:\/\/multai.eu\/wp-content\/uploads\/2024\/10\/Logo.png\",\"width\":225,\"height\":244,\"caption\":\"WSW\"},\"image\":{\"@id\":\"https:\/\/multai.eu\/#\/schema\/logo\/image\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/multai.eu\/#\/schema\/person\/06def8c374b5d6724bec911e9880c292\",\"name\":\"hans\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"de\",\"@id\":\"https:\/\/multai.eu\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/1409f6643b6f17d5838709af9deca41643884a95390f8a4f8ea478b9187aec41?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/1409f6643b6f17d5838709af9deca41643884a95390f8a4f8ea478b9187aec41?s=96&d=mm&r=g\",\"caption\":\"hans\"},\"sameAs\":[\"https:\/\/wsw-int.de\"],\"url\":\"https:\/\/multai.eu\/de\/author\/hans\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Techniques for using language models - MultAI","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/multai.eu\/de\/techniques-for-using-language-models\/","og_locale":"de_DE","og_type":"article","og_title":"Techniques for using language models - MultAI","og_description":"One of the weaknesses of the models currently available on the market is that they have been trained on a publicly accessible data set, which may not necessarily be sufficient to meet certain specific needs. Take, for example, a company with a large volume of proprietary data, a highly specialized vocabulary or specific data formats. [&hellip;]","og_url":"https:\/\/multai.eu\/de\/techniques-for-using-language-models\/","og_site_name":"MultAI","article_published_time":"2024-03-24T13:47:31+00:00","article_modified_time":"2024-10-25T13:24:15+00:00","og_image":[{"url":"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp"}],"author":"hans","twitter_card":"summary_large_image","twitter_misc":{"Verfasst von":"hans","Gesch\u00e4tzte Lesezeit":"7\u00a0Minuten"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#article","isPartOf":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/"},"author":{"name":"hans","@id":"https:\/\/multai.eu\/#\/schema\/person\/06def8c374b5d6724bec911e9880c292"},"headline":"Techniques for using language models","datePublished":"2024-03-24T13:47:31+00:00","dateModified":"2024-10-25T13:24:15+00:00","mainEntityOfPage":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/"},"wordCount":1429,"commentCount":0,"publisher":{"@id":"https:\/\/multai.eu\/#organization"},"image":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp","keywords":["finetuning","GenAI","LLM","prompt","prompting","RAG","retrieval-augmented generation"],"articleSection":["Uncategorized"],"inLanguage":"de","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/multai.eu\/techniques-for-using-language-models\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/","url":"https:\/\/multai.eu\/techniques-for-using-language-models\/","name":"Techniques for using language models - MultAI","isPartOf":{"@id":"https:\/\/multai.eu\/#website"},"primaryImageOfPage":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage"},"image":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage"},"thumbnailUrl":"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp","datePublished":"2024-03-24T13:47:31+00:00","dateModified":"2024-10-25T13:24:15+00:00","breadcrumb":{"@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#breadcrumb"},"inLanguage":"de","potentialAction":[{"@type":"ReadAction","target":["https:\/\/multai.eu\/techniques-for-using-language-models\/"]}]},{"@type":"ImageObject","inLanguage":"de","@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#primaryimage","url":"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp","contentUrl":"https:\/\/wsw-int.de\/wp-content\/uploads\/2024\/03\/image-20240218a-1024x578.webp"},{"@type":"BreadcrumbList","@id":"https:\/\/multai.eu\/techniques-for-using-language-models\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/multai.eu\/"},{"@type":"ListItem","position":2,"name":"Techniques for using language models"}]},{"@type":"WebSite","@id":"https:\/\/multai.eu\/#website","url":"https:\/\/multai.eu\/","name":"WSW","description":"Generative AI for your business","publisher":{"@id":"https:\/\/multai.eu\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/multai.eu\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"de"},{"@type":"Organization","@id":"https:\/\/multai.eu\/#organization","name":"WSW","alternateName":"MultAI","url":"https:\/\/multai.eu\/","logo":{"@type":"ImageObject","inLanguage":"de","@id":"https:\/\/multai.eu\/#\/schema\/logo\/image\/","url":"https:\/\/multai.eu\/wp-content\/uploads\/2024\/10\/Logo.png","contentUrl":"https:\/\/multai.eu\/wp-content\/uploads\/2024\/10\/Logo.png","width":225,"height":244,"caption":"WSW"},"image":{"@id":"https:\/\/multai.eu\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/multai.eu\/#\/schema\/person\/06def8c374b5d6724bec911e9880c292","name":"hans","image":{"@type":"ImageObject","inLanguage":"de","@id":"https:\/\/multai.eu\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/1409f6643b6f17d5838709af9deca41643884a95390f8a4f8ea478b9187aec41?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/1409f6643b6f17d5838709af9deca41643884a95390f8a4f8ea478b9187aec41?s=96&d=mm&r=g","caption":"hans"},"sameAs":["https:\/\/wsw-int.de"],"url":"https:\/\/multai.eu\/de\/author\/hans\/"}]}},"_links":{"self":[{"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/posts\/115","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/comments?post=115"}],"version-history":[{"count":3,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/posts\/115\/revisions"}],"predecessor-version":[{"id":1426,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/posts\/115\/revisions\/1426"}],"wp:attachment":[{"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/media?parent=115"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/categories?post=115"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/multai.eu\/de\/wp-json\/wp\/v2\/tags?post=115"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}