事前学習済み言語モデル
原題: Pretrained language model
分析結果
- カテゴリ
- AI
- 重要度
- 78
- トレンドスコア
- 42
- 要約
- 生成的事前学習トランスフォーマー(GPT)は、生成的人工知能チャットボットで広く使用されている大規模言語モデル(LLM)の一種です。GPTは、トランスフォーマーと呼ばれる深層学習アーキテクチャに基づいています。
- キーワード
A generative pre-trained transformer (GPT) is a type of large language model (LLM) that is widely used in generative artificial intelligence chatbots. GPTs are based on a deep learning architecture called the transformer. They are pre-trained on large datasets of unlabeled content, and able to generate novel content.OpenAI was the first to apply generative pre-training to the transformer architecture, introducing the GPT-1 model in 2018. The company has since released many bigger GPT models. The chatbot ChatGPT, released in late 2022 (using GPT-3.5), was followed by many competitor chatbots using their own generative pre-trained transformers to generate text, such as Gemini, DeepSeek and Claude.GPTs are primarily used to generate text, but can be trained to generate other kinds of data. For example, GPT-4o can process and generate text, images and audio. To improve performance on complex tasks, some GPTs, such as OpenAI o3, allocate more computation time analyzing the problem before generating an output, and are called reasoning models. In 2025, GPT-5 was released with a router that automatically selects whether to use a faster model or slower reasoning model based on the provided task.