Meta Llama 2

開源商用的 Meta Llama 2 (70B / 13B / 7B) 全系列模型

Meta Llama 2

★ 提供 70B / 13B / 7B 選項 ★
★ 使用 2 兆 token 訓練的資料量 ★
★ 支援達 4,096 個 token 的上下文長度 ★

Applicable Scenarios

在推理、編碼能力和知識測試等方面的基準測試中優於其他開源大語言模型

70B: Excellent understanding and reasoning skills in structured data, tables, and Markdown; ability to join data tables and use SQL and JSON; and proficiency in various application scenarios of 13B and 7B.

13B: Suitable for Markdown intent analysis + JSON format output, and various applicable scenarios of 7B.

7B: Can be used for marketing slogans and content generation, invitation and email writing, Chinese-English translation, article summarization, de-identification, and chatbot Q&A.

Parameter adjustment

Standard

Full parameter fine-tuning involves fine-tuning all parameters of a pre-trained model using a task-specific dataset, which requires significant GPU computing and storage resources.

PEFT

参数高效微调(Parameter-Efficient Fine-Tuning),固定大部分預訓練參數,僅微調少量額外的模型參數,降低了運算和儲存成本,同時趨近與全量參數微調的效果 (可適用的方法為LoRA)。

LoRA (Low-Rank Adaptation of Large Language Models) fixes the pre-trained model parameters and weights, and adds additional network layer modules. By using a high-dimensional to low-dimensional decomposition matrix module, it updates only the parameters of the network layer modules to simulate the process of full parameter fine-tuning. This significantly reduces the number of parameters required for training, reduces computational and storage resources, and achieves indirect training of large models with fewer parameters, while approaching the effect of full model fine-tuning.

Free Consultation Service

Contact Taiwan AI Cloud experts to learn about and start using the solution that suits you.

Customer Service & Technical Support

EDM Subscription

On-Demand AI Cloud Consulting

EDM Subscription

Sales Contact
Sales Contact Form