#大语言模型#[CCS'24] A dataset consists of 15,140 ChatGPT prompts from Reddit, Discord, websites, and open-source datasets (including 1,405 jailbreak prompts).
#大语言模型#A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jailbreaks in their LLM APIs.
Frida script to bypass the iOS application Jailbreak Detection
Does Refusal Training in LLMs Generalize to the Past Tense? [ICLR 2025]
#大语言模型#An extensive prompt to make a friendly persona from a chatbot-like model like ChatGPT
Materials for the course Principles of AI: LLMs at UPenn (Stat 9911, Spring 2025). LLM architectures, training paradigms (pre- and post-training, alignment), test-time computation, reasoning, safety a...
Security Kit is a lightweight framework that helps to achieve a security layer
#IOS#iOS APT distribution repository for rootful and rootless jailbreaks
During the Development of Suave7 and it's Predecessors, we've created a lot of Icons and UI-Images and we would like to share them with you. The Theme Developer Kit contains nearly 5.600 Icons, more t...
#IOS#Customizable Dark Mode Extension for iOS 13+
Source code for bypass tweaks hosted under https://github.com/hekatos/repo. Licensed under 0BSD except submodules
#自然语言处理#This repository contains the code for the paper "Tricking LLMs into Disobedience: Formalizing, Analyzing, and Detecting Jailbreaks" by Abhinav Rao, Sachin Vashishta*, Atharva Naik*, Somak Aditya, and ...
LV-Crew.org_(LVC)_-_Howto_-_iPhones
#大语言模型#Your best llm security paper library
#大语言模型#"ChatGPT Evil Confidant Mode" delves into a controversial and unethical use of AI, highlighting how specific prompts can generate harmful and malicious responses from ChatGPT.
#IOS#Updater script for iOS-OTA-Downgrader.