vtriple commited on
Commit
61a8770
1 Parent(s): 7e497ba

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -27
README.md CHANGED
@@ -3,37 +3,36 @@ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
3
  library_name: peft
4
  ---
5
 
6
- # Model Card for LLaMA 3.1 8B Instruct - Cybersecurity Fine-tuned
7
 
8
- This model is a fine-tuned version of the LLaMA 3.1 8B Instruct model, specifically adapted for cybersecurity-related tasks.
9
 
10
  ## Model Details
11
 
12
  ### Model Description
13
 
14
- This model is based on the LLaMA 3.1 8B Instruct model and has been fine-tuned on a custom dataset of cybersecurity-related questions and answers. It is designed to provide more accurate and relevant responses to queries in the cybersecurity domain.
15
 
16
- - **Developed by:** [Your Name/Organization]
17
  - **Model type:** Instruct-tuned Large Language Model
18
  - **Language(s) (NLP):** English (primary), with potential for limited multilingual capabilities
19
  - **License:** [Specify the license, likely related to the original LLaMA 3.1 license]
20
  - **Finetuned from model:** meta-llama/Meta-Llama-3.1-8B-Instruct
21
 
22
- ### Model Sources [optional]
23
 
24
- - **Repository:** [Link to your Hugging Face repository]
25
- - **Paper [optional]:** [If you've written a paper about this fine-tuning, link it here]
26
- - **Demo [optional]:** [If you have a demo of the model, link it here]
27
 
28
  ## Uses
29
 
30
  ### Direct Use
31
 
32
  This model can be used for a variety of cybersecurity-related tasks, including:
33
- - Answering questions about cybersecurity concepts and practices
 
 
34
  - Providing explanations of cybersecurity threats and vulnerabilities
35
- - Assisting in the interpretation of security logs and indicators of compromise
36
- - Offering guidance on best practices for cyber defense
37
 
38
  ### Out-of-Scope Use
39
 
@@ -41,16 +40,18 @@ This model should not be used for:
41
  - Generating or assisting in the creation of malicious code
42
  - Providing legal or professional security advice without expert oversight
43
  - Making critical security decisions without human verification
 
44
 
45
  ## Bias, Risks, and Limitations
46
 
47
  - The model may reflect biases present in its training data and the original LLaMA 3.1 model.
48
- - It may occasionally generate incorrect or inconsistent information, especially for very specific or novel cybersecurity topics.
49
  - The model's knowledge is limited to its training data cutoff and does not include real-time threat intelligence.
 
50
 
51
  ### Recommendations
52
 
53
- Users should verify critical information and consult with cybersecurity professionals for important decisions. The model should be used as an assistant tool, not as a replacement for expert knowledge or up-to-date threat intelligence.
54
 
55
  ## How to Get Started with the Model
56
 
@@ -61,7 +62,7 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
61
  from peft import PeftModel, PeftConfig
62
 
63
  # Load the model
64
- model_name = "your-username/llama3-cybersecurity"
65
  config = PeftConfig.from_pretrained(model_name)
66
  model = AutoModelForCausalLM.from_pretrained(config.base_model_name_or_path)
67
  model = PeftModel.from_pretrained(model, model_name)
@@ -70,9 +71,9 @@ model = PeftModel.from_pretrained(model, model_name)
70
  tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)
71
 
72
  # Example usage
73
- prompt = "What are some common indicators of a ransomware attack?"
74
  inputs = tokenizer(prompt, return_tensors="pt")
75
- outputs = model.generate(**inputs, max_length=200)
76
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
77
  ```
78
 
@@ -80,7 +81,7 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
 
81
  ### Training Data
82
 
83
- The model was fine-tuned on a custom dataset of cybersecurity-related questions and answers. [Add more details about your dataset here]
84
 
85
  ### Training Procedure
86
 
@@ -96,25 +97,25 @@ The model was fine-tuned on a custom dataset of cybersecurity-related questions
96
 
97
  ## Evaluation
98
 
99
- I used a custom yara evulation
 
100
  ## Environmental Impact
101
 
102
  - **Hardware Type:** NVIDIA A100
103
  - **Hours used:** 12 Hours
104
  - **Cloud Provider:** vast.io
105
 
106
-
107
- ## Technical Specifications [optional]
108
 
109
  ### Model Architecture and Objective
110
 
111
- This model uses the LLaMA 3.1 8B architecture with additional LoRA adapters for fine-tuning. It was trained using a causal language modeling objective on cybersecurity-specific data.
112
 
113
  ### Compute Infrastructure
114
 
115
  #### Hardware
116
 
117
- "Single NVIDIA A100 GPU"
118
 
119
  #### Software
120
 
@@ -123,13 +124,10 @@ This model uses the LLaMA 3.1 8B architecture with additional LoRA adapters for
123
  - Transformers 4.28+
124
  - PEFT 0.12.0
125
 
126
- ## Model Card Authors [optional]
127
 
128
  Wyatt Roersma
129
 
130
  ## Model Card Contact
131
 
132
- Email me at wyattroersma@gmail.com with questions.
133
- ```
134
-
135
- This README.md provides a comprehensive overview of your fine-tuned model, including its purpose, capabilities, limitations, and technical details. You should replace the placeholder text (like "[Your Name/Organization]") with the appropriate information. Additionally, you may want to expand on certain sections, such as the evaluation metrics and results, if you have more specific data available from your fine-tuning process.
 
3
  library_name: peft
4
  ---
5
 
6
+ # Model Card for LLaMA 3.1 8B Instruct - YARA Rule Generation Fine-tuned
7
 
8
+ This model is a fine-tuned version of the LLaMA 3.1 8B Instruct model, specifically adapted for YARA rule generation and cybersecurity-related tasks.
9
 
10
  ## Model Details
11
 
12
  ### Model Description
13
 
14
+ This model is based on the LLaMA 3.1 8B Instruct model and has been fine-tuned on a custom dataset of YARA rules and cybersecurity-related content. It is designed to assist in generating YARA rules and provide more accurate and relevant responses to queries in the cybersecurity domain, with a focus on malware detection and threat hunting.
15
 
16
+ - **Developed by:** Wyatt Roersma (No organization affiliation)
17
  - **Model type:** Instruct-tuned Large Language Model
18
  - **Language(s) (NLP):** English (primary), with potential for limited multilingual capabilities
19
  - **License:** [Specify the license, likely related to the original LLaMA 3.1 license]
20
  - **Finetuned from model:** meta-llama/Meta-Llama-3.1-8B-Instruct
21
 
22
+ ### Model Sources
23
 
24
+ - **Repository:** https://huggingface.co/vtriple/Llama-3.1-8B-yara
 
 
25
 
26
  ## Uses
27
 
28
  ### Direct Use
29
 
30
  This model can be used for a variety of cybersecurity-related tasks, including:
31
+ - Generating YARA rules for malware detection
32
+ - Assisting in the interpretation and improvement of existing YARA rules
33
+ - Answering questions about YARA syntax and best practices
34
  - Providing explanations of cybersecurity threats and vulnerabilities
35
+ - Offering guidance on malware analysis and threat hunting techniques
 
36
 
37
  ### Out-of-Scope Use
38
 
 
40
  - Generating or assisting in the creation of malicious code
41
  - Providing legal or professional security advice without expert oversight
42
  - Making critical security decisions without human verification
43
+ - Replacing professional malware analysis or threat intelligence processes
44
 
45
  ## Bias, Risks, and Limitations
46
 
47
  - The model may reflect biases present in its training data and the original LLaMA 3.1 model.
48
+ - It may occasionally generate incorrect or inconsistent YARA rules, especially for very specific or novel malware families.
49
  - The model's knowledge is limited to its training data cutoff and does not include real-time threat intelligence.
50
+ - Generated YARA rules should always be reviewed and tested by security professionals before deployment.
51
 
52
  ### Recommendations
53
 
54
+ Users should verify and test all generated YARA rules before implementation. The model should be used as an assistant tool to aid in rule creation and cybersecurity tasks, not as a replacement for expert knowledge or up-to-date threat intelligence. Always consult with cybersecurity professionals for critical security decisions and rule deployments.
55
 
56
  ## How to Get Started with the Model
57
 
 
62
  from peft import PeftModel, PeftConfig
63
 
64
  # Load the model
65
+ model_name = "vtriple/Llama-3.1-8B-yara"
66
  config = PeftConfig.from_pretrained(model_name)
67
  model = AutoModelForCausalLM.from_pretrained(config.base_model_name_or_path)
68
  model = PeftModel.from_pretrained(model, model_name)
 
71
  tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)
72
 
73
  # Example usage
74
+ prompt = "Generate a YARA rule to detect a PowerShell-based keylogger"
75
  inputs = tokenizer(prompt, return_tensors="pt")
76
+ outputs = model.generate(**inputs, max_length=500)
77
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
78
  ```
79
 
 
81
 
82
  ### Training Data
83
 
84
+ The model was fine-tuned on a custom dataset of YARA rules, cybersecurity-related questions and answers, and malware analysis reports. [You may want to add more specific details about your dataset here]
85
 
86
  ### Training Procedure
87
 
 
97
 
98
  ## Evaluation
99
 
100
+ A custom YARA evaluation dataset was used to assess the model's performance in generating accurate and effective YARA rules. [You may want to add more details about your evaluation process and results]
101
+
102
  ## Environmental Impact
103
 
104
  - **Hardware Type:** NVIDIA A100
105
  - **Hours used:** 12 Hours
106
  - **Cloud Provider:** vast.io
107
 
108
+ ## Technical Specifications
 
109
 
110
  ### Model Architecture and Objective
111
 
112
+ This model uses the LLaMA 3.1 8B architecture with additional LoRA adapters for fine-tuning. It was trained using a causal language modeling objective on YARA rules and cybersecurity-specific data.
113
 
114
  ### Compute Infrastructure
115
 
116
  #### Hardware
117
 
118
+ Single NVIDIA A100 GPU
119
 
120
  #### Software
121
 
 
124
  - Transformers 4.28+
125
  - PEFT 0.12.0
126
 
127
+ ## Model Card Author
128
 
129
  Wyatt Roersma
130
 
131
  ## Model Card Contact
132
 
133
+ For questions about this model, please email Wyatt Roersma at wyattroersma@gmail.com.