<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>devkuma – LLM</title>
    <link>https://www.devkuma.com/en/tags/llm/</link>
    <image>
      <url>https://www.devkuma.com/en/tags/llm/logo/180x180.jpg</url>
      <title>LLM</title>
      <link>https://www.devkuma.com/en/tags/llm/</link>
    </image>
    <description>Recent content in LLM on devkuma</description>
    <generator>Hugo -- gohugo.io</generator>
    <language>en</language>
    <managingEditor>kc@example.com (kc kim)</managingEditor>
    <webMaster>kc@example.com (kc kim)</webMaster>
    <copyright>The devkuma</copyright>
    
	  <atom:link href="https://www.devkuma.com/en/tags/llm/index.xml" rel="self" type="application/rss+xml" />
    
    
      
        
      
    
    
    <item>
      <title>LLM (Large Language Model)</title>
      <link>https://www.devkuma.com/en/docs/ai/llm/</link>
      <pubDate>Sun, 24 Aug 2025 13:14:00 +0900</pubDate>
      <author>kc@example.com (kc kim)</author>
      <guid>https://www.devkuma.com/en/docs/ai/llm/</guid>
      <description>
        
        
        &lt;h2 id=&#34;llm-overview&#34;&gt;LLM Overview&lt;/h2&gt;
&lt;p&gt;LLMs (Large Language Models) are &lt;strong&gt;artificial intelligence models that learn from massive amounts of text data and can understand and generate natural language&lt;/strong&gt;. They mainly use a &lt;strong&gt;deep learning-based transformer architecture&lt;/strong&gt;, so they statistically capture the characteristics of human language and have advanced text generation and processing capabilities.&lt;/p&gt;
&lt;p&gt;LLMs are now a central part of AI and play a very important role in language-based applications and system design.&lt;/p&gt;
&lt;h2 id=&#34;how-llms-work&#34;&gt;How LLMs Work&lt;/h2&gt;
&lt;h3 id=&#34;learning-method-and-transformer-architecture&#34;&gt;Learning Method and Transformer Architecture&lt;/h3&gt;
&lt;p&gt;LLMs perform pre-training through &lt;strong&gt;unsupervised learning&lt;/strong&gt; on hundreds of billions of text examples.&lt;br&gt;
In particular, the &lt;strong&gt;transformer architecture&lt;/strong&gt; understands contextual relationships through self-attention, and because it can process data in parallel compared with earlier recurrent neural networks (RNNs), it is highly efficient for training.&lt;/p&gt;
&lt;h3 id=&#34;parameters-and-embeddings&#34;&gt;Parameters and Embeddings&lt;/h3&gt;
&lt;p&gt;The term &amp;ldquo;large&amp;rdquo; refers to the size of the parameters, which can range from billions to hundreds of billions. These enormous parameters make it possible to capture complex contexts and nuances in language.
In addition, an &amp;ldquo;embedding&amp;rdquo; converts words into multidimensional vectors and numerically represents semantic similarity, helping the model understand context.&lt;/p&gt;
&lt;h2 id=&#34;application-areas&#34;&gt;Application Areas&lt;/h2&gt;
&lt;p&gt;LLMs can be used very flexibly. Representative applications include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Generative AI&lt;/strong&gt;: Generates text such as essays, translations, and summaries based on user prompts&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Code generation&lt;/strong&gt;: Supports code writing from natural language, as seen in GitHub Copilot, AWS CodeWhisperer, and similar tools&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Text classification and sentiment analysis&lt;/strong&gt;: Customer feedback classification, document clustering, and more&lt;/li&gt;
&lt;li&gt;Others: Knowledge-based question answering (KI-NLP), chatbots, customer service automation, and more&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;types-of-learning-methods&#34;&gt;Types of Learning Methods&lt;/h2&gt;
&lt;p&gt;There are three main ways to use an LLM for a specific purpose:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Zero-shot learning&lt;/strong&gt;: Performs various tasks with general prompts without additional training&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Few-shot learning&lt;/strong&gt;: Improves performance by providing a small number of examples&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Fine-tuning&lt;/strong&gt;: Further trains parameters on specific data to enable specialized use&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;importance-and-expected-benefits&#34;&gt;Importance and Expected Benefits&lt;/h2&gt;
&lt;p&gt;Adopting LLMs can bring various benefits to companies and organizations:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Work automation&lt;/strong&gt;: Improves productivity by automating language-based tasks such as customer support, document summarization, and content generation&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Scalability and flexibility&lt;/strong&gt;: A single model can flexibly handle multiple tasks such as translation, summarization, and question answering&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Encouraging innovation&lt;/strong&gt;: Provides a foundation for many future possibilities, including knowledge extraction, creative assistance, and conversational interfaces&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;limitations-and-considerations&#34;&gt;Limitations and Considerations&lt;/h2&gt;
&lt;p&gt;When using LLMs, the following limitations should also be considered:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;High resource requirements&lt;/strong&gt;: Training and serving models with billions of parameters requires substantial computing resources.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Potential bias and errors&lt;/strong&gt;: Limitations or biases in training data can be reflected in model outputs, requiring continuous improvement in accuracy.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Privacy and security concerns&lt;/strong&gt;: Systems must prepare for possible relationships with private or sensitive data.&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h2 id=&#34;summary&#34;&gt;Summary&lt;/h2&gt;
&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Item&lt;/th&gt;
          &lt;th&gt;Description&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;Definition&lt;/td&gt;
          &lt;td&gt;A massive text-based deep learning model capable of natural language understanding and generation&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;How it works&lt;/td&gt;
          &lt;td&gt;Transformer-based, with self-attention, embeddings, and billions of parameters&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Applications&lt;/td&gt;
          &lt;td&gt;Text generation, code generation, classification, summarization, chatbots, and more&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Learning methods&lt;/td&gt;
          &lt;td&gt;Zero-shot, few-shot, fine-tuning&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Advantages&lt;/td&gt;
          &lt;td&gt;Automation, scalability, and creative use&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Limitations&lt;/td&gt;
          &lt;td&gt;Resource demands, bias and accuracy issues, security risks, and more&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;

      </description>
      
      <category>AI</category>
      
      <category>ChatGPT</category>
      
      <category>LLM</category>
      
    </item>
    
    <item>
      <title>Multi-Model</title>
      <link>https://www.devkuma.com/en/docs/ai/multi-model/</link>
      <pubDate>Sat, 30 Aug 2025 13:14:00 +0900</pubDate>
      <author>kc@example.com (kc kim)</author>
      <guid>https://www.devkuma.com/en/docs/ai/multi-model/</guid>
      <description>
        
        
        &lt;h2 id=&#34;what-is-multi-model&#34;&gt;What Is Multi-Model?&lt;/h2&gt;
&lt;p&gt;It refers to &lt;strong&gt;an approach that uses multiple models together in a single AI system&lt;/strong&gt;.&lt;br&gt;
In other words, instead of assigning everything to a single model, it &lt;strong&gt;combines the strengths of each model&lt;/strong&gt; to achieve better performance or more diverse functions.&lt;/p&gt;
&lt;p&gt;For example, it may be a model that can process not only text but also images, audio, and video together.&lt;/p&gt;
&lt;h2 id=&#34;why-is-it-needed&#34;&gt;Why Is It Needed?&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;When one model is not enough&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Example: When both images and text must be handled&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Use of specialized models&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Uses a large general-purpose model together with domain-specific models&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Performance optimization&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Heavy and slow models are used only for core reasoning, while lightweight models handle preprocessing and simple tasks&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Cost reduction&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Always using a huge model like GPT-4 is expensive, so some tasks are assigned to smaller models and only difficult parts use a larger model&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;h2 id=&#34;types-of-multi-model&#34;&gt;Types of Multi-Model&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Different from Multi-Modal&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Multi-Model != Multi-Modal&lt;/li&gt;
&lt;li&gt;&lt;em&gt;Multi-Modal&lt;/em&gt;: One model that processes &lt;strong&gt;multiple input forms&lt;/strong&gt;, such as images, text, and speech&lt;/li&gt;
&lt;li&gt;&lt;em&gt;Multi-Model&lt;/em&gt;: A system built by &lt;strong&gt;combining multiple models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Configuration methods&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Parallel (Ensemble)&lt;/strong&gt;: Multiple models produce answers at the same time, and the results are combined to make the final decision&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Examples: Voting, blending, weighted sum&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Serial (Pipeline)&lt;/strong&gt;: The output of one model is passed as the input to another model&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Example: Image captioning model -&amp;gt; text summarization model -&amp;gt; question answering model&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Hybrid&lt;/strong&gt;: Selects models depending on the situation, such as with a router model&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;h2 id=&#34;examples&#34;&gt;Examples&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Retrieval + generation (RAG)&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;Retrieval model (vector search) + generative model (LLM)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Copilot-style tools&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;Code assistance: a small model for fast code completion, GPT-4 for sophisticated bug fixes&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Autonomous driving&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;Video recognition CNN + behavior planning RL model&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Healthcare&lt;/strong&gt;
&lt;ul&gt;
&lt;li&gt;Medical knowledge model + general LLM combination&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;multi-model-vs-single-model&#34;&gt;Multi-Model vs Single Model&lt;/h2&gt;
&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Category&lt;/th&gt;
          &lt;th&gt;Single Model&lt;/th&gt;
          &lt;th&gt;Multi-Model&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;strong&gt;Structure&lt;/strong&gt;&lt;/td&gt;
          &lt;td&gt;One model performs everything&lt;/td&gt;
          &lt;td&gt;Multiple models divide roles&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;/td&gt;
          &lt;td&gt;Simple and easy to manage&lt;/td&gt;
          &lt;td&gt;Higher accuracy, more flexibility, and ability to use the latest technologies&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;/td&gt;
          &lt;td&gt;General-purpose models have performance limits&lt;/td&gt;
          &lt;td&gt;System is complex and requires coordination&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;h2 id=&#34;summary&#34;&gt;Summary&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Multi-Model is a system design approach that combines multiple models and uses each model&amp;rsquo;s strengths to produce better results&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Examples include combining a &amp;ldquo;retrieval model + generative model,&amp;rdquo; &amp;ldquo;small model + large model,&amp;rdquo; or &amp;ldquo;specialized model + general-purpose model.&amp;rdquo;&lt;/p&gt;

      </description>
      
      <category>AI</category>
      
      <category>ChatGPT</category>
      
      <category>LLM</category>
      
    </item>
    
  </channel>
</rss>
