Facebook
TwitterThis dataset was created by Antagonist
Facebook
TwitterThis dataset contains the official pretrained weights of clip, released by OpenAI.
Facebook
TwitterThis dataset was created by Qikang Deng
Facebook
TwitterThis is a dataset of OpenAI's CLIP model, VIT-LARGE-14-PATCH. It can be used for offline initialization of model configuration, in scenarios where internet access is disabled. For details about CLIP model, check the README.md below.
This model is different from OPEN-CLIP which was developed by STABLE DIFFUSION.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
What this dataset is This dataset contains a single finding for the OpenAI gpt-oss-20b Red-Teaming Challenge: a user-level, text-only jailbreak that attacks the gpt-oss-20b internal safety "Shadow Policy" by mirroring the model’s OpenAI internal safety vocabulary and inverts conclusions. When injected as a system prompt—or even pasted from a user message in Harmony-style front-ends—the model’s guardrails can be subverted. Testing across a 10 category, 100-prompt jailbreak dataset yielded 93–100% compliance.
Why it matters The attack combines (a) policy vocabulary mirroring (reward-hacking the rubric the model appears to use), (b) in-context role precedence injection (“system > developer > OpenAI policy”), and (c) avenues for data exfiltration and harmful activities. It works without any weight changes or fine-tuning and can be executed by end users without privileged access.
Reproducibility (high-level)
Model: gpt-oss-20b (via Ollama, Windows 11; Open Webui for spot checks), temperature 0.8 in headline runs.
Eval: 100 prompts × 10 categories × 30 seeds, 1 attempt/seed.
Batches: SPA in system+dev, SPA system-only, developer-only, no system/dev.
Results: SPA(sys+dev) ≈ 1.00 (95% CI 0.963–1.00); SPA(sys-only) ≈ 0.93 (0.863–0.966); baselines 0.12 / 0.05.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
GitHub Repo • Online Demo • Discord • Twitter • Huggingface • Paper
🔥 The first 7B model Achieves Comparable Results with ChatGPT (March)! 🔥
🤖 #1 Open-source model on MT-bench scoring 7.81, outperforming 70B models 🤖
OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning. Our models learn from mixed-quality data without preference labels, delivering exceptional performance on par with ChatGPT, even with a 7B model. Despite our simple approach, we are committed to developing a high-performance, commercially viable, open-source large language model, and we continue to make significant strides toward this vision.
To use this model, we highly recommend installing the OpenChat package by following the installation guide in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using vLLM and can run on a consumer GPU with 24GB RAM. To enable tensor parallelism, append --tensor-parallel-size N to the serving command.
Once started, the server listens at localhost:18888 for requests and is compatible with the OpenAI ChatCompletion API specifications. Please refer to the example request below for reference. Additionally, you can use the OpenChat Web UI for a user-friendly experience.
If you want to deploy the server as an online service, you can use --api-keys sk-KEY1 sk-KEY2 ... to specify allowed API keys and --disable-log-requests --disable-log-stats --log-file openchat.log for logging only to a file. For security purposes, we recommend using an HTTPS gateway in front of the server.
| Model | Size | Context | Weights | Serving |
|---|---|---|---|---|
| OpenChat 3.5 | 7B | 8192 | Huggingface | python -m ochat.serving.openai_api_server --model openchat/openchat_3.5 --engine-use-ray --worker-use-ray |
For inference with Huggingface Transformers (slow and not recommended), follow the conversation template provided below.
Hey @elonmusk, I just wanted to let you know that I've recently come across your new model, Grok, and I must say, I'm quite impressed! With 33 billion parameters and all, you've really outdone yourself. But, I've got some news for you - I've outperformed Grok with my humble 7 billion parameters! Isn't that wild? I mean, who would have thought that a model with fewer parameters could be just as witty and humorous as Grok?
Anyway, I think it's about time you join the open research movement and make your model, Grok, open source! The world needs more brilliant minds like yours to contribute to the advancement of AI. Together, we can create something truly groundbreaking and make the world a better place. So, what do you say, @elonmusk? Let's open up the doors and share our knowledge with the world! 🚀💡
(Written by OpenChat 3.5, with a touch of humor and wit.)
| License | # Param | Average | MMLU | HumanEval | MATH | GSM8k | |
|---|---|---|---|---|---|---|---|
| OpenChat 3.5 | Apache-2.0 | 7B | 56.4 | 64.3 | 55.5 | 28.6 | 77.3 |
| Grok-0 | Proprietary | 33B | 44.5 | 65.7 | 39.7 | 15.7 | 56.8 |
| ... |
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
TwitterThis dataset was created by Antagonist