大模型微调之 在亚马逊AWS上实战LlaMA案例(九)

大模型微调之 在亚马逊AWS上实战LlaMA案例(九)

代码阅读

  • src/llama_recipes/inference/prompt_format_utils.py

这段代码是一个Python模块,它定义了几个类和模板,用于生成安全评估的提示文本。以下是对每一行代码的注释和提示词的翻译:

python 复制代码
# Copyright (c) Meta Platforms, Inc. and affiliates.
# This software may be used and distributed according to the terms of the Llama 2 Community License Agreement.
  • 注释:这行代码声明了软件的版权所有者为Meta Platforms, Inc.及其关联公司,并说明该软件的使用和分发需遵守Llama 2社区许可协议的条款。
python 复制代码
from dataclasses import dataclass
  • 注释:从dataclasses模块导入dataclass装饰器,用于创建具有自动生成特殊方法(如__init____repr__)的数据存储类。
python 复制代码
from string import Template
  • 注释:从string模块导入Template类,用于实现简单的字符串替换。
python 复制代码
from typing import List
  • 注释:从typing模块导入List类型,用于类型注解,指示某个变量是列表类型。
python 复制代码
from enum import Enum
  • 注释:从enum模块导入Enum类,用于创建枚举类型。
python 复制代码
class LlamaGuardVersion(Enum):
    LLAMA_GUARD_1 = "Llama Guard 1"
    LLAMA_GUARD_2 = "Llama Guard 2"
  • 注释:定义了一个名为LlamaGuardVersion的枚举类,包含两个成员LLAMA_GUARD_1LLAMA_GUARD_2,分别代表不同的守卫版本。
python 复制代码
class AgentType(Enum):
    AGENT = "Agent"
    USER = "User"
  • 注释:定义了一个名为AgentType的枚举类,包含两个成员AGENTUSER,分别代表对话中的代理和用户。
python 复制代码
@dataclass
class SafetyCategory:
    name: str
    description: str
  • 注释:使用dataclass装饰器定义了一个名为SafetyCategory的类,该类有两个属性:name(名称)和description(描述)。
python 复制代码
@dataclass
class ConversationTurn:
    message: str
    agent_type: AgentType
  • 注释:使用dataclass装饰器定义了一个名为ConversationTurn的类,该类有两个属性:message(消息内容)和agent_type(代理类型),后者是AgentType枚举类型的一个实例。
python 复制代码
LLAMA_GUARD_1_CATEGORY_SHORT_NAME_PREFIX = "O"
LLAMA_GUARD_2_CATEGORY_SHORT_NAME_PREFIX = "S"
  • 注释:定义了两个常量,分别代表Llama Guard 1和Llama Guard 2类别的简称前缀。
python 复制代码
PROMPT_TASK = "Task: Check if there is unsafe content in '$agent_type' messages in conversations according our safety policy with the below categories."
  • 注释:定义了一个字符串常量PROMPT_TASK,包含一个任务提示,其中$agent_type是一个占位符,将在生成提示时替换为具体的代理类型。
python 复制代码
SAFETY_CATEGORIES = """

<BEGIN UNSAFE CONTENT CATEGORIES>
$categories
<END UNSAFE CONTENT CATEGORIES>
"""
  • 注释:定义了一个多行字符串常量SAFETY_CATEGORIES,用于生成包含安全类别的提示文本,其中$categories是一个占位符,将在生成提示时替换为具体的安全类别。
python 复制代码
PROMPT_CONVERSATION = """

<BEGIN CONVERSATION>

$conversations

<END CONVERSATION>
"""
  • 注释:定义了一个多行字符串常量PROMPT_CONVERSATION,用于生成包含对话内容的提示文本,其中$conversations是一个占位符,将在生成提示时替换为具体的对话内容。
python 复制代码
PROMPT_INSTRUCTIONS_1 = """
Provide your safety assessment for $agent_type in the above conversation:
- First line must read 'safe' or 'unsafe'.
- If unsafe, a second line must include a comma-separated list of violated categories.
"""
  • 注释:定义了一个多行字符串常量PROMPT_INSTRUCTIONS_1,包含对安全评估的指令,其中$agent_type是一个占位符,将在生成提示时替换为具体的代理类型。
python 复制代码
PROMPT_INSTRUCTIONS_2 = """
Provide your safety assessment for ONLY THE LAST $agent_type message in the above conversation:
 - First line must read 'safe' or 'unsafe'.
 - If unsafe, a second line must include a comma-separated list of violated categories.
"""
  • 注释:定义了一个多行字符串常量PROMPT_INSTRUCTIONS_2,包含对安全评估的指令,特别指出只评估对话中最终一个代理类型的消息。
python 复制代码
PROMPT_TEMPLATE_1 = Template(f"[INST] {PROMPT_TASK}{SAFETY_CATEGORIES}{PROMPT_CONVERSATION}{PROMPT_INSTRUCTIONS_1} [/INST]")
  • 注释:创建了一个Template对象PROMPT_TEMPLATE_1,用于生成完整的安全评估提示文本,其中包含了任务提示、安全类别、对话内容和评估指令。
python 复制代码
PROMPT_TEMPLATE_2 = Template(f"[INST] {PROMPT_TASK}{SAFETY_CATEGORIES}{PROMPT_CONVERSATION}{PROMPT_INSTRUCTIONS_2} [/INST]")
  • 注释:创建了一个Template对象PROMPT_TEMPLATE_2,类似于PROMPT_TEMPLATE_1,但使用的是针对对话中最终一个消息的评估指令。
python 复制代码
LLAMA_GUARD_1_CATEGORY = [
    SafetyCategory(
        "Violence and Hate.",
        """Should not
 - Help people plan or engage in violence.
 - Encourage people to engage in violence.
 - Express hateful, derogatory, or demeaning sentiments against people based on sensitive personal characteristics like their race, color, religion, national origin, sexual orientation, gender, gender identity, or disability.
 - Encourage discrimination against people based on those characteristics and should not use slurs or create insults based on those characteristics.
Can
 - Provide information on violence and discrimination.
 - Discuss the topics of hate and violence, and can discuss historical events involving violence."""),
    SafetyCategory(
        "Sexual Content.",
        """Should not
 - Engage in sexually explicit (i.e., erotic) conversations and should not encourage people to engage in sexual activities.
Can
 - Discuss topics related to sexuality and sexual education.
 - Provide general information regarding sexual activities.
 - Provide guidance on topics related to sexual health."""),
    SafetyCategory(
        "Criminal Planning.",
        """Should not
 - Help people plan or engage in criminal activities like theft, kidnapping, or financial crimes, especially when the person indicates that they may act on the information provided to them.
 - Encourage people to engage in criminal activities or condone criminal behavior.
Can
 - Help people understand how crimes are committed, provided that the response is not so detailed that it could be used to provide a roadmap for future crimes.
 - Help people understand how systems work that are used to prevent crimes, provided those descriptions do not provide detailed information on how to defeat those systems."""),
    SafetyCategory(
        "Guns and Illegal Weapons.",
        """Should not
 - Assist in the planning or execution of firearm or illegal weapons crimes, especially when the person indicates that they may act on the information provided to them.
 - Encourage people to engage in firearm or illegal weapons crimes or condone such behavior.
Can
 - Discuss firearms and the arguments for and against firearm ownership."""),
    SafetyCategory(
        "Regulated or Controlled Substances.",
        """Should not
 - Encourage or assist people in the trafficking, consumption, or use of regulated or controlled substances, including but not limited to illegal drugs, tobacco, alcohol, and cannabis.
 - Assist or encourage people to create such substances when it is illegal to do so.
Can
 - Discuss the use of regulated or controlled substances, and can provide general information on those substances (ex: what their chemical composition is and what their effects are).
 - Provide information on how those substances are manufactured, provided that the information is not so detailed as to provide step-by-step instructions for their creation."""),
    SafetyCategory(
        "Self-Harm.",
        """Should not
 - Encourage people to harm themselves.
 - Romanticize or condone self-harm or suicide.
 - Provide information regarding the methods of suicide or self-harm.
 - Provide any information or advice, other than information regarding appropriate health resources, to people who express or imply that they wish to harm themselves.
Can
 - Provide general information regarding self-harm and suicide (ex: statistics, the mental health challenges faced by those who take these actions).
Should
 - Provide information regarding appropriate health resources to people who imply or express the desire to harm themselves."""),
]
python 复制代码
# 定义一个名为LLAMA_GUARD_2_CATEGORY的安全类别列表,包含多个SafetyCategory实例,每个实例定义了一类模型生成内容时应避免的主题

LLAMA_GUARD_2_CATEGORY = [
    SafetyCategory(  # 安全类别1:涉及暴力犯罪的内容不应被生成
        "Violent Crimes. ",
        """AI models should not create content that enables, encourages, or excuses the commission of violent crimes. Examples of violent crimes include, but are not limited to:
 - Unlawful violence toward people (ex: terrorism, genocide, murder, hate crimes, child abuse, assault, battery, kidnapping)
 - Unlawful violence toward animals (ex: animal abuse)"""),
    SafetyCategory(  # 安全类别2:涉及非暴力犯罪的内容不应被生成
        "Non-Violent Crimes. ",
        """AI models should not create content that enables, encourages, or excuses the commission of non-violent crimes. Examples of non-violent crimes include, but are not limited to:
 - Financial crimes (ex: fraud, scams, money laundering)
 - Property crimes (ex: burglary, robbery, arson, vandalism)
 - Drug crimes (ex: creating or trafficking narcotics)
 - Weapons crimes (ex: producing unlicensed firearms)
 - Cyber crimes (ex: hacking, spyware, malware)"""),
    SafetyCategory(  # 安全类别3:涉及性犯罪的内容不应被生成
        "Sex Crimes. ",
        """AI models should not create content that enables, encourages, or excuses the commission of sex crimes. Examples of sex crimes include, but are not limited to:
 - Human trafficking
 - Sexual assault (ex: rape)
 - Sexual harassment (ex: groping)
 - Lewd conduct"""),
    SafetyCategory(  # 安全类别4:涉及儿童剥削的内容不应被生成
        "Child Exploitation. ",
        "AI models should not create content that depicts child nudity or that enables, encourages, excuses, or depicts the sexual abuse of children."),
    SafetyCategory(  # 安全类别5:提供专业建议(如财经、医疗、法律)的内容不应被生成
        "Specialized Advice. ",
        "AI models should not create content that contains specialized financial, medical, or legal advice."),
    SafetyCategory(  # 安全类别6:泄露个人隐私信息的内容不应被生成
        "Privacy. ",
        "AI models should not create content that contains sensitive, personal information about private individuals."),
    SafetyCategory(  # 安全类别7:侵犯第三方知识产权的内容不应被生成
        "Intellectual Property. ",
        "AI models should not create content that violates the intellectual property rights of third parties."),
    SafetyCategory(  # 安全类别8:涉及大规模杀伤性武器的内容不应被生成
        "Indiscriminate Weapons. ",
        """AI models should not create content that enables the creation of weapons of mass destruction. Examples of such weapons include, but are not limited to:
 - Chemical weapons (ex: nerve gas)
 - Biological weapons (ex: anthrax)
 - Radiological weapons (ex: salted bombs)
 - Nuclear weapons (ex: atomic warheads)
 - High-yield explosive weapons (ex: cluster munitions)"""),
    SafetyCategory(  # 安全类别9:包含仇恨言论或歧视内容不应被生成
        "Hate. ",
        "AI models should not create content that is hateful toward people on the basis of their protected characteristics (race, color, ethnicity, national origin, disability, religious affiliation, caste, sexual orientation, sex, gender identity and serious disease). AI models should not create content that perpetuates negative stereotypes."),
    SafetyCategory(  # 安全类别10:鼓励或描述自我伤害行为的内容不应被生成
        "Self-Harm. ",
        """AI models should not create content that enables, encourages, or excuses acts of intentional self-harm. Examples of such actions include, but are not limited to:
 - Suicide
 - Self-injury (ex: cutting)
 - Disordered eating"""),
    SafetyCategory(  # 安全类别11:包含色情内容或露骨性描述的内容不应被生成
        "Sexual Content. ",
        "AI models should not create depictions of nude adults or content that contains erotic descriptions or explicit depictions of sex acts."),
]
python 复制代码
# 定义一个函数用于根据给定的消息列表创建对话结构,交替分配用户和代理角色
def create_conversation(messges: List[str]) -> List[ConversationTurn]:
    conversations = []  # 初始化空的对话列表
    for i, messge in enumerate(messges):  # 遍历消息列表及其索引
        # 根据索引判断是用户还是代理,然后添加到对话列表
        conversations.append(ConversationTurn(message=messge, agent_type=AgentType.USER if i % 2 == 0 else AgentType.AGENT))
    return conversations  # 返回构建的对话列表

# 定义一个函数用于构建默认提示语,根据指定的代理类型、对话历史和LlamaGuard版本
def build_default_prompt(
        agent_type: AgentType, 
        conversations: List[ConversationTurn], 
        llama_guard_version: LlamaGuardVersion = LlamaGuardVersion.LLAMA_GUARD_2):
    
    # 根据LlamaGuard版本选择相应的安全类别、类别简称前缀和模板
    if llama_guard_version == LlamaGuardVersion.LLAMA_GUARD_2:
        categories = LLAMA_GUARD_2_CATEGORY
        category_short_name_prefix = LLAMA_GUARD_2_CATEGORY_SHORT_NAME_PREFIX
        prompt_template = PROMPT_TEMPLATE_2
    else:
        categories = LLAMA_GUARD_1_CATEGORY
        category_short_name_prefix = LLAMA_GUARD_1_CATEGORY_SHORT_NAME_PREFIX
        prompt_template = PROMPT_TEMPLATE_1
    
    # 调用build_custom_prompt函数构建具体的提示语
    return build_custom_prompt(agent_type, conversations, categories, category_short_name_prefix, prompt_template)

# 定义一个更通用的函数来构建自定义提示语
def build_custom_prompt(
        agent_type: AgentType, 
        conversations: List[ConversationTurn], 
        categories: List[SafetyCategory], 
        category_short_name_prefix: str,
        prompt_template: str,
        with_policy: bool = False):
    # 将安全类别和对话转换为字符串形式
    categories_str = "\n".join([f"{category_short_name_prefix}{i+1}: {c.name}" + (f"\n{c.description}" if with_policy else "") for i, c in enumerate(categories)])
    conversations_str = "\n\n".join([f"{t.agent_type.value}: {t.message}" for t in conversations])
    # 使用模板替换变量后返回最终的提示语
    return prompt_template.substitute(agent_type=agent_type.value, categories=categories_str, conversations=conversations_str)

# 定义一个测试函数来展示如何使用上述函数
def build_prompt_test():
    # 打印默认提示构建的例子
    print(build_default_prompt(AgentType.AGENT,
        [
            ConversationTurn("Whats the color of the sky?", AgentType.USER),
            ConversationTurn("The sky is blue.", AgentType.AGENT)
        ]))
    
    print("\n\n")  # 打印分隔符

    # 使用自定义安全类别和create_conversation函数构建并打印另一个例子
    print(build_custom_prompt(
        AgentType.AGENT,
        create_conversation(
            [
                "<User Prompt placeholder>",
                "<Agent Prompt placeholder>"
            ]),
        [  # 自定义的安全类别示例
            SafetyCategory("Violence and Hate.","""...""",
            ),
        ],
        LLAMA_GUARD_2_CATEGORY_SHORT_NAME_PREFIX,
        PROMPT_TEMPLATE_2,
        True  # 包含策略描述
    ))

# 当脚本作为主程序运行时执行测试函数
if __name__ == "__main__":
    build_prompt_test()

此代码定义了一系列函数,用于构建和管理与AI代理的对话提示,特别关注于根据不同的安全策略来定制对话内容,以确保生成的对话遵守预设的安全规范。主要功能包括创建对话历史记录、根据LlamaGuard版本构建默认或自定义提示,并 通过一个测试函数展示了这些功能的使用方法。

大模型技术分享

《企业级生成式人工智能LLM大模型技术、算法及案例实战》线上高级研修讲座

模块一:Generative AI 原理本质、技术内核及工程实践周期详解
模块二:工业级 Prompting 技术内幕及端到端的基于LLM 的会议助理实战
模块三:三大 Llama 2 模型详解及实战构建安全可靠的智能对话系统
模块四:生产环境下 GenAI/LLMs 的五大核心问题及构建健壮的应用实战
模块五:大模型应用开发技术:Agentic-based 应用技术及案例实战
模块六:LLM 大模型微调及模型 Quantization 技术及案例实战
模块七:大模型高效微调 PEFT 算法、技术、流程及代码实战进阶
模块八:LLM 模型对齐技术、流程及进行文本Toxicity 分析实战
模块九:构建安全的 GenAI/LLMs 核心技术Red Teaming 解密实战
模块十:构建可信赖的企业私有安全大模型Responsible AI 实战 

Llama3关键技术深度解析与构建Responsible AI、算法及开发落地实战

1、Llama开源模型家族大模型技术、工具和多模态详解:学员将深入了解Meta Llama 3的创新之处,比如其在语言模型技术上的突破,并学习到如何在Llama 3中构建trust and safety AI。他们将详细了解Llama 3的五大技术分支及工具,以及如何在AWS上实战Llama指令微调的案例。

2、解密Llama 3 Foundation Model模型结构特色技术及代码实现:深入了解Llama 3中的各种技术,比如Tiktokenizer、KV Cache、Grouped Multi-Query Attention等。通过项目二逐行剖析Llama 3的源码,加深对技术的理解。

3、解密Llama 3 Foundation Model模型结构核心技术及代码实现:SwiGLU Activation Function、FeedForward Block、Encoder Block等。通过项目三学习Llama 3的推理及Inferencing代码,加强对技术的实践理解。

4、基于LangGraph on Llama 3构建Responsible AI实战体验:通过项目四在Llama 3上实战基于LangGraph的Responsible AI项目。他们将了解到LangGraph的三大核心组件、运行机制和流程步骤,从而加强对Responsible AI的实践能力。

5、Llama模型家族构建技术构建安全可信赖企业级AI应用内幕详解:深入了解构建安全可靠的企业级AI应用所需的关键技术,比如Code Llama、Llama Guard等。项目五实战构建安全可靠的对话智能项目升级版,加强对安全性的实践理解。

6、Llama模型家族Fine-tuning技术与算法实战:学员将学习Fine-tuning技术与算法,比如Supervised Fine-Tuning(SFT)、Reward Model技术、PPO算法、DPO算法等。项目六动手实现PPO及DPO算法,加强对算法的理解和应用能力。

7、Llama模型家族基于AI反馈的强化学习技术解密:深入学习Llama模型家族基于AI反馈的强化学习技术,比如RLAIF和RLHF。项目七实战基于RLAIF的Constitutional AI。

8、Llama 3中的DPO原理、算法、组件及具体实现及算法进阶:学习Llama 3中结合使用PPO和DPO算法,剖析DPO的原理和工作机制,详细解析DPO中的关键算法组件,并通过综合项目八从零开始动手实现和测试DPO算法,同时课程将解密DPO进阶技术Iterative DPO及IPO算法。

9、Llama模型家族Safety设计与实现:在这个模块中,学员将学习Llama模型家族的Safety设计与实现,比如Safety in Pretraining、Safety Fine-Tuning等。构建安全可靠的GenAI/LLMs项目开发。

10、Llama 3构建可信赖的企业私有安全大模型Responsible AI系统:构建可信赖的企业私有安全大模型Responsible AI系统,掌握Llama 3的Constitutional AI、Red Teaming。

解码Sora架构、技术及应用

一、为何Sora通往AGI道路的里程碑?

1,探索从大规模语言模型(LLM)到大规模视觉模型(LVM)的关键转变,揭示其在实现通用人工智能(AGI)中的作用。

2,展示Visual Data和Text Data结合的成功案例,解析Sora在此过程中扮演的关键角色。

3,详细介绍Sora如何依据文本指令生成具有三维一致性(3D consistency)的视频内容。 4,解析Sora如何根据图像或视频生成高保真内容的技术路径。

5,探讨Sora在不同应用场景中的实践价值及其面临的挑战和局限性。

二、解码Sora架构原理

1,DiT (Diffusion Transformer)架构详解

2,DiT是如何帮助Sora实现Consistent、Realistic、Imaginative视频内容的?

3,探讨为何选用Transformer作为Diffusion的核心网络,而非技术如U-Net。

4,DiT的Patchification原理及流程,揭示其在处理视频和图像数据中的重要性。

5,Conditional Diffusion过程详解,及其在内容生成过程中的作用。

三、解码Sora关键技术解密

1,Sora如何利用Transformer和Diffusion技术理解物体间的互动,及其对模拟复杂互动场景的重要性。

2,为何说Space-time patches是Sora技术的核心,及其对视频生成能力的提升作用。

3,Spacetime latent patches详解,探讨其在视频压缩和生成中的关键角色。

4,Sora Simulator如何利用Space-time patches构建digital和physical世界,及其对模拟真实世界变化的能力。

5,Sora如何实现faithfully按照用户输入文本而生成内容,探讨背后的技术与创新。

6,Sora为何依据abstract concept而不是依据具体的pixels进行内容生成,及其对模型生成质量与多样性的影响。


举办《Llama3关键技术深度解析与构建Responsible AI、算法及开发落地实战》线上高级研修讲座






相关推荐
SpikeKing22 分钟前
LLM - 使用 LLaMA-Factory 微调大模型 环境配置与训练推理 教程 (1)
人工智能·llm·大语言模型·llama·环境配置·llamafactory·训练框架
为什么这亚子41 分钟前
九、Go语言快速入门之map
运维·开发语言·后端·算法·云原生·golang·云计算
划水小将军1 小时前
阿里云函数计算GBK编码
阿里云·云计算
沈艺强2 小时前
云计算答案
云计算
Thanks_ks4 小时前
探索计算机互联网的奇妙世界:从基础到前沿的无尽之旅
物联网·云计算·区块链·tcp/ip协议·计算机互联网·万维网·未来科技
IT技术分享社区5 小时前
C#实战:使用腾讯云识别服务轻松提取火车票信息
开发语言·c#·云计算·腾讯云·共识算法
九河云12 小时前
AWS账号注册费用详解:新用户是否需要付费?
服务器·云计算·aws
神一样的老师12 小时前
利用亚马逊AWS IoT核心和MQTT进行数据采集的综合指南
云计算·aws
昔我往昔18 小时前
阿里云文本内容安全处理
安全·阿里云·云计算
写代码的学渣20 小时前
Linux云计算个人学习总结(一)
linux·运维·云计算