LF

LLM-Tuning-Safety/LLMs-Finetuning-Safety

We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20 via OpenAI’s APIs.

345 36 +0/wk
GitHub
alignment llm llm-finetuning
Trend 0

Star & Fork Trend (18 data points)

Stars
Forks

Multi-Source Signals

Growth Velocity

LLM-Tuning-Safety/LLMs-Finetuning-Safety has +0 stars this period . Velocity data will be available after more historical data is collected.

Deep analysis is being generated for this repository.

Signal-backed technical analysis will be available soon.

Metric LLMs-Finetuning-Safety Build-a-Large-Language-Model-from-Scratch advanced-rag agent
Stars 345 345343347
Forks 36 8414118
Weekly Growth +0 +0+0+1
Language Python PythonJupyter NotebookTypeScript
Sources 1 111
License MIT N/AMITMIT

Capability Radar vs Build-a-Large-Language-Model-from-Scratch

LLMs-Finetuning-Safety
Build-a-Large-Language-Model-from-Scratch
Maintenance Activity 0

Last code push 775 days ago.

Community Engagement 52

Fork-to-star ratio: 10.4%. Active community forking and contributing.

Issue Burden 70

Issue data not yet available.

Growth Momentum 30

No measurable growth in the current period (first-day cold start expected).

License Clarity 95

Licensed under MIT. Permissive — safe for commercial use.

Risk scores are computed from real-time repository data. Higher scores indicate healthier metrics.