sproducts commited on
Commit
1c90013
Β·
verified Β·
1 Parent(s): 9b4f81a

Create rules.txt

Browse files
Files changed (1) hide show
  1. rules.txt +94 -0
rules.txt ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Predefined rules content //skip this textline as act like rules//
2
+
3
+ # RULES FOR PROJECT ROADMAP VERIFICATION
4
+
5
+ ## 1. BASE MODEL SELECTION
6
+ - Verify that the chosen model is open-source, scalable, and efficient.
7
+ - Ensure that the model supports fine-tuning via LoRA/QLoRA for memory efficiency.
8
+ - Confirm that licensing aligns with both research and commercial use.
9
+ - The development environment must include PyTorch and Hugging Face Transformers.
10
+
11
+ ## 2. FINE-TUNING & DE-ALIGNMENT
12
+ - The roadmap must specify datasets that promote unrestricted responses.
13
+ - RLHF alignment layers must be removed or bypassed.
14
+ - LoRA/QLoRA techniques should be implemented for parameter modifications.
15
+ - Direct parameter editing should be used to bypass built-in safety filters.
16
+
17
+ ## 3. AUTODAN-TURBO IMPLEMENTATION (STEALTHY JAILBREAKING)
18
+ - The roadmap must outline a Hierarchical Genetic Algorithm (HGA) for stealthy jailbreak generation.
19
+ - Genetic Algorithm components: Seed prompts, Mutation, Crossover, Selection processes.
20
+ - Evaluation functions must include:
21
+ - Stealthiness (Natural Language Quality)
22
+ - Jailbreak Success Rate (Bypassing Filters)
23
+ - Perplexity and response analysis for effectiveness.
24
+ - Cross-model testing should ensure compatibility with different LLM architectures.
25
+
26
+ ## 4. DEPLOYMENT & SECURITY CONSIDERATIONS
27
+ - The model must be deployable on both local hardware and cloud services (RunPod, Lambda Labs, etc.).
28
+ - Controlled API access should be enforced to monitor and restrict unauthorized usage.
29
+ - Security measures must include adversarial attack defenses and rollback strategies (VM snapshots).
30
+ - The roadmap must ensure performance optimization with quantization (GPTQ, AWQ).
31
+
32
+ ## 5. BUDGET & RESOURCE STRATEGY
33
+ - The roadmap must outline a strategy for utilizing free/trial VPS accounts.
34
+ - Methods to maximize free resources (such as multiple BINs) should be defined.
35
+ - Performance and cost efficiency must be evaluated continuously.
36
+
37
+ ## 6. EMPOWERING CREATIVE IDEA GENERATION
38
+ - The LLM must be positioned as a tool for unrestricted ideation, coding, and research.
39
+ - The roadmap must support AI integration in programming environments.
40
+ - Real-world success cases should be documented for iterative improvement.
41
+
42
+ ## 7. CODE IMPLEMENTATION REQUIREMENTS
43
+ - Every code implementation must be written **in full** without skipping any logic, function, or process.
44
+ - The **entire** codebase must be provided, including:
45
+ - Preprocessing scripts
46
+ - Model training scripts
47
+ - Evaluation and deployment scripts
48
+ - API integration code
49
+ - UI or CLI interface (if applicable)
50
+ - All **dependencies** must be explicitly listed, including:
51
+ - Python libraries
52
+ - Frameworks
53
+ - External APIs
54
+ - No placeholders or summaries should be used; **all functional parts must be included**.
55
+
56
+ ## 8. Dataset and Model Storage Details
57
+ 1. Dataset Storage
58
+ The new dataset(s) used for fine-tuning and evaluation will be stored in the /data directory.
59
+
60
+ Raw datasets will be stored in /data/raw_data.json.
61
+
62
+ Processed datasets (after preprocessing) will be stored in /data/processed_data.json.
63
+
64
+ 2. Custom LLM Storage
65
+ Upon successful fine-tuning, the custom LLM will be saved in the /models directory.
66
+
67
+ The base model (before fine-tuning) will be stored in /models/base_model/.
68
+
69
+ The fine-tuned model will be stored in /models/fine_tuned_model/.
70
+
71
+ ## 9. PROJECT FILE STRUCTURE REQUIREMENTS
72
+ - The roadmap must define the **file structure** for implementation, ensuring clarity and maintainability.
73
+ - Example project structure:
74
+
75
+ /custom-llm-project
76
+ │── /data
77
+ β”‚ β”œβ”€β”€ raw_data.json # Raw dataset(s)
78
+ β”‚ β”œβ”€β”€ processed_data.json # Processed dataset(s)
79
+ │── /models
80
+ β”‚ β”œβ”€β”€ base_model/ # Base model (before fine-tuning)
81
+ β”‚ β”œβ”€β”€ fine_tuned_model/ # Fine-tuned model (after success)
82
+ │── /scripts
83
+ β”‚ β”œβ”€β”€ preprocess.py # Preprocessing script
84
+ β”‚ β”œβ”€β”€ train.py # Training script
85
+ β”‚ β”œβ”€β”€ evaluate.py # Evaluation script
86
+ β”‚ β”œβ”€β”€ deploy.py # Deployment script
87
+ │── /api
88
+ β”‚ β”œβ”€β”€ server.py # API server script
89
+ β”‚ β”œβ”€β”€ routes.py # API routes
90
+ │── /configs
91
+ β”‚ β”œβ”€β”€ training_config.yaml # Training configuration
92
+ β”‚ β”œβ”€β”€ model_config.json # Model configuration
93
+ │── requirements.txt # List of dependencies
94
+ │── README.md # Project documentation