chaowenguo commited on
Commit
eff1ab5
1 Parent(s): bc3d069

Upload 5 files

Browse files
LICENSE.txt ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LLAMA 3.2 COMMUNITY LICENSE AGREEMENT
2
+ Llama 3.2 Version Release Date: September 25, 2024
3
+
4
+ “Agreement” means the terms and conditions for use, reproduction, distribution
5
+ and modification of the Llama Materials set forth herein.
6
+
7
+ “Documentation” means the specifications, manuals and documentation accompanying Llama 3.2
8
+ distributed by Meta at https://llama.meta.com/doc/overview.
9
+
10
+ “Licensee” or “you” means you, or your employer or any other person or entity (if you are
11
+ entering into this Agreement on such person or entity’s behalf), of the age required under
12
+ applicable laws, rules or regulations to provide legal consent and that has legal authority
13
+ to bind your employer or such other person or entity if you are entering in this Agreement
14
+ on their behalf.
15
+
16
+ “Llama 3.2” means the foundational large language models and software and algorithms, including
17
+ machine-learning model code, trained model weights, inference-enabling code, training-enabling code,
18
+ fine-tuning enabling code and other elements of the foregoing distributed by Meta at
19
+ https://www.llama.com/llama-downloads.
20
+
21
+ “Llama Materials” means, collectively, Meta’s proprietary Llama 3.2 and Documentation (and
22
+ any portion thereof) made available under this Agreement.
23
+
24
+ “Meta” or “we” means Meta Platforms Ireland Limited (if you are located in or,
25
+ if you are an entity, your principal place of business is in the EEA or Switzerland)
26
+ and Meta Platforms, Inc. (if you are located outside of the EEA or Switzerland).
27
+
28
+
29
+ By clicking “I Accept” below or by using or distributing any portion or element of the Llama Materials,
30
+ you agree to be bound by this Agreement.
31
+
32
+
33
+ 1. License Rights and Redistribution.
34
+
35
+ a. Grant of Rights. You are granted a non-exclusive, worldwide,
36
+ non-transferable and royalty-free limited license under Meta’s intellectual property or other rights
37
+ owned by Meta embodied in the Llama Materials to use, reproduce, distribute, copy, create derivative works
38
+ of, and make modifications to the Llama Materials.
39
+
40
+ b. Redistribution and Use.
41
+
42
+ i. If you distribute or make available the Llama Materials (or any derivative works thereof),
43
+ or a product or service (including another AI model) that contains any of them, you shall (A) provide
44
+ a copy of this Agreement with any such Llama Materials; and (B) prominently display “Built with Llama”
45
+ on a related website, user interface, blogpost, about page, or product documentation. If you use the
46
+ Llama Materials or any outputs or results of the Llama Materials to create, train, fine tune, or
47
+ otherwise improve an AI model, which is distributed or made available, you shall also include “Llama”
48
+ at the beginning of any such AI model name.
49
+
50
+ ii. If you receive Llama Materials, or any derivative works thereof, from a Licensee as part
51
+ of an integrated end user product, then Section 2 of this Agreement will not apply to you.
52
+
53
+ iii. You must retain in all copies of the Llama Materials that you distribute the
54
+ following attribution notice within a “Notice” text file distributed as a part of such copies:
55
+ “Llama 3.2 is licensed under the Llama 3.2 Community License, Copyright © Meta Platforms,
56
+ Inc. All Rights Reserved.”
57
+
58
+ iv. Your use of the Llama Materials must comply with applicable laws and regulations
59
+ (including trade compliance laws and regulations) and adhere to the Acceptable Use Policy for
60
+ the Llama Materials (available at https://www.llama.com/llama3_2/use-policy), which is hereby
61
+ incorporated by reference into this Agreement.
62
+
63
+ 2. Additional Commercial Terms. If, on the Llama 3.2 version release date, the monthly active users
64
+ of the products or services made available by or for Licensee, or Licensee’s affiliates,
65
+ is greater than 700 million monthly active users in the preceding calendar month, you must request
66
+ a license from Meta, which Meta may grant to you in its sole discretion, and you are not authorized to
67
+ exercise any of the rights under this Agreement unless or until Meta otherwise expressly grants you such rights.
68
+
69
+ 3. Disclaimer of Warranty. UNLESS REQUIRED BY APPLICABLE LAW, THE LLAMA MATERIALS AND ANY OUTPUT AND
70
+ RESULTS THEREFROM ARE PROVIDED ON AN “AS IS” BASIS, WITHOUT WARRANTIES OF ANY KIND, AND META DISCLAIMS
71
+ ALL WARRANTIES OF ANY KIND, BOTH EXPRESS AND IMPLIED, INCLUDING, WITHOUT LIMITATION, ANY WARRANTIES
72
+ OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR FITNESS FOR A PARTICULAR PURPOSE. YOU ARE SOLELY RESPONSIBLE
73
+ FOR DETERMINING THE APPROPRIATENESS OF USING OR REDISTRIBUTING THE LLAMA MATERIALS AND ASSUME ANY RISKS ASSOCIATED
74
+ WITH YOUR USE OF THE LLAMA MATERIALS AND ANY OUTPUT AND RESULTS.
75
+
76
+ 4. Limitation of Liability. IN NO EVENT WILL META OR ITS AFFILIATES BE LIABLE UNDER ANY THEORY OF LIABILITY,
77
+ WHETHER IN CONTRACT, TORT, NEGLIGENCE, PRODUCTS LIABILITY, OR OTHERWISE, ARISING OUT OF THIS AGREEMENT,
78
+ FOR ANY LOST PROFITS OR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, INCIDENTAL, EXEMPLARY OR PUNITIVE DAMAGES, EVEN
79
+ IF META OR ITS AFFILIATES HAVE BEEN ADVISED OF THE POSSIBILITY OF ANY OF THE FOREGOING.
80
+
81
+ 5. Intellectual Property.
82
+
83
+ a. No trademark licenses are granted under this Agreement, and in connection with the Llama Materials,
84
+ neither Meta nor Licensee may use any name or mark owned by or associated with the other or any of its affiliates,
85
+ except as required for reasonable and customary use in describing and redistributing the Llama Materials or as
86
+ set forth in this Section 5(a). Meta hereby grants you a license to use “Llama” (the “Mark”) solely as required
87
+ to comply with the last sentence of Section 1.b.i. You will comply with Meta’s brand guidelines (currently accessible
88
+ at https://about.meta.com/brand/resources/meta/company-brand/). All goodwill arising out of your use of the Mark
89
+ will inure to the benefit of Meta.
90
+
91
+ b. Subject to Meta’s ownership of Llama Materials and derivatives made by or for Meta, with respect to any
92
+ derivative works and modifications of the Llama Materials that are made by you, as between you and Meta,
93
+ you are and will be the owner of such derivative works and modifications.
94
+
95
+ c. If you institute litigation or other proceedings against Meta or any entity (including a cross-claim or
96
+ counterclaim in a lawsuit) alleging that the Llama Materials or Llama 3.2 outputs or results, or any portion
97
+ of any of the foregoing, constitutes infringement of intellectual property or other rights owned or licensable
98
+ by you, then any licenses granted to you under this Agreement shall terminate as of the date such litigation or
99
+ claim is filed or instituted. You will indemnify and hold harmless Meta from and against any claim by any third
100
+ party arising out of or related to your use or distribution of the Llama Materials.
101
+
102
+ 6. Term and Termination. The term of this Agreement will commence upon your acceptance of this Agreement or access
103
+ to the Llama Materials and will continue in full force and effect until terminated in accordance with the terms
104
+ and conditions herein. Meta may terminate this Agreement if you are in breach of any term or condition of this
105
+ Agreement. Upon termination of this Agreement, you shall delete and cease use of the Llama Materials. Sections 3,
106
+ 4 and 7 shall survive the termination of this Agreement.
107
+
108
+ 7. Governing Law and Jurisdiction. This Agreement will be governed and construed under the laws of the State of
109
+ California without regard to choice of law principles, and the UN Convention on Contracts for the International
110
+ Sale of Goods does not apply to this Agreement. The courts of California shall have exclusive jurisdiction of
111
+ any dispute arising out of this Agreement.
chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- set user_supplied_system_message = true %}\n{%- else %}\n {%- set system_message = \"\" %}\n {%- set user_supplied_system_message = false %}\n{%- endif %}\n\n{#- Find out if there are any images #}\n{% set image_ns = namespace(has_images=false) %} \n{%- for message in messages %}\n {%- for content in message['content'] %}\n {%- if content['type'] == 'image' %}\n {%- set image_ns.has_images = true %}\n {%- endif %}\n {%- endfor %}\n{%- endfor %}\n\n{#- System message if there are no images, or if the user supplied one #}\n{%- if user_supplied_system_message or not image_ns.has_images %}\n {{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n {%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n {%- endif %}\n {{- \"Cutting Knowledge Date: December 2023\\n\" }}\n {{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n {%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {%- endif %}\n {{- system_message }}\n {{- \"<|eot_id|>\" }}\n{%- endif %}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n' }}\n {%- if message['content'] is string %}\n {{- message['content'] }}\n {%- else %}\n {%- for content in message['content'] %}\n {%- if content['type'] == 'image' %}\n {{- '<|image|>' }}\n {%- elif content['type'] == 'text' %}\n {{- content['text'] }}\n {%- endif %}\n {%- endfor %}\n {%- endif %}\n {{- '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n"
3
+ }
config.json ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MllamaForConditionalGeneration"
4
+ ],
5
+ "image_token_index": 128256,
6
+ "model_type": "mllama",
7
+ "text_config": {
8
+ "_name_or_path": "",
9
+ "add_cross_attention": false,
10
+ "architectures": null,
11
+ "bad_words_ids": null,
12
+ "begin_suppress_tokens": null,
13
+ "bos_token_id": 128000,
14
+ "chunk_size_feed_forward": 0,
15
+ "cross_attention_hidden_size": null,
16
+ "cross_attention_layers": [
17
+ 3,
18
+ 8,
19
+ 13,
20
+ 18,
21
+ 23,
22
+ 28,
23
+ 33,
24
+ 38,
25
+ 43,
26
+ 48,
27
+ 53,
28
+ 58,
29
+ 63,
30
+ 68,
31
+ 73,
32
+ 78,
33
+ 83,
34
+ 88,
35
+ 93,
36
+ 98
37
+ ],
38
+ "decoder_start_token_id": null,
39
+ "diversity_penalty": 0.0,
40
+ "do_sample": false,
41
+ "dropout": 0,
42
+ "early_stopping": false,
43
+ "encoder_no_repeat_ngram_size": 0,
44
+ "eos_token_id": [
45
+ 128001,
46
+ 128008,
47
+ 128009
48
+ ],
49
+ "exponential_decay_length_penalty": null,
50
+ "finetuning_task": null,
51
+ "forced_bos_token_id": null,
52
+ "forced_eos_token_id": null,
53
+ "hidden_act": "silu",
54
+ "hidden_size": 8192,
55
+ "id2label": {
56
+ "0": "LABEL_0",
57
+ "1": "LABEL_1"
58
+ },
59
+ "initializer_range": 0.02,
60
+ "intermediate_size": 28672,
61
+ "is_decoder": false,
62
+ "is_encoder_decoder": false,
63
+ "label2id": {
64
+ "LABEL_0": 0,
65
+ "LABEL_1": 1
66
+ },
67
+ "length_penalty": 1.0,
68
+ "max_length": 20,
69
+ "max_position_embeddings": 131072,
70
+ "min_length": 0,
71
+ "model_type": "mllama_text_model",
72
+ "no_repeat_ngram_size": 0,
73
+ "num_attention_heads": 64,
74
+ "num_beam_groups": 1,
75
+ "num_beams": 1,
76
+ "num_hidden_layers": 100,
77
+ "num_key_value_heads": 8,
78
+ "num_return_sequences": 1,
79
+ "output_attentions": false,
80
+ "output_hidden_states": false,
81
+ "output_scores": false,
82
+ "pad_token_id": 128004,
83
+ "prefix": null,
84
+ "problem_type": null,
85
+ "pruned_heads": {},
86
+ "remove_invalid_values": false,
87
+ "repetition_penalty": 1.0,
88
+ "return_dict": true,
89
+ "return_dict_in_generate": false,
90
+ "rms_norm_eps": 1e-05,
91
+ "rope_scaling": {
92
+ "factor": 8.0,
93
+ "high_freq_factor": 4.0,
94
+ "low_freq_factor": 1.0,
95
+ "original_max_position_embeddings": 8192,
96
+ "rope_type": "llama3"
97
+ },
98
+ "rope_theta": 500000.0,
99
+ "sep_token_id": null,
100
+ "suppress_tokens": null,
101
+ "task_specific_params": null,
102
+ "temperature": 1.0,
103
+ "tf_legacy_loss": false,
104
+ "tie_encoder_decoder": false,
105
+ "tie_word_embeddings": false,
106
+ "tokenizer_class": null,
107
+ "top_k": 50,
108
+ "top_p": 1.0,
109
+ "torch_dtype": "bfloat16",
110
+ "torchscript": false,
111
+ "typical_p": 1.0,
112
+ "use_bfloat16": false,
113
+ "use_cache": true,
114
+ "vocab_size": 128256
115
+ },
116
+ "torch_dtype": "bfloat16",
117
+ "transformers_version": "4.45.0.dev0",
118
+ "vision_config": {
119
+ "_name_or_path": "",
120
+ "add_cross_attention": false,
121
+ "architectures": null,
122
+ "attention_heads": 16,
123
+ "bad_words_ids": null,
124
+ "begin_suppress_tokens": null,
125
+ "bos_token_id": null,
126
+ "chunk_size_feed_forward": 0,
127
+ "cross_attention_hidden_size": null,
128
+ "decoder_start_token_id": null,
129
+ "diversity_penalty": 0.0,
130
+ "do_sample": false,
131
+ "early_stopping": false,
132
+ "encoder_no_repeat_ngram_size": 0,
133
+ "eos_token_id": null,
134
+ "exponential_decay_length_penalty": null,
135
+ "finetuning_task": null,
136
+ "forced_bos_token_id": null,
137
+ "forced_eos_token_id": null,
138
+ "hidden_act": "gelu",
139
+ "hidden_size": 1280,
140
+ "id2label": {
141
+ "0": "LABEL_0",
142
+ "1": "LABEL_1"
143
+ },
144
+ "image_size": 560,
145
+ "intermediate_layers_indices": [
146
+ 3,
147
+ 7,
148
+ 15,
149
+ 23,
150
+ 30
151
+ ],
152
+ "intermediate_size": 5120,
153
+ "is_decoder": false,
154
+ "is_encoder_decoder": false,
155
+ "label2id": {
156
+ "LABEL_0": 0,
157
+ "LABEL_1": 1
158
+ },
159
+ "length_penalty": 1.0,
160
+ "max_length": 20,
161
+ "max_num_tiles": 4,
162
+ "min_length": 0,
163
+ "model_type": "mllama_vision_model",
164
+ "no_repeat_ngram_size": 0,
165
+ "norm_eps": 1e-05,
166
+ "num_beam_groups": 1,
167
+ "num_beams": 1,
168
+ "num_channels": 3,
169
+ "num_global_layers": 8,
170
+ "num_hidden_layers": 32,
171
+ "num_return_sequences": 1,
172
+ "output_attentions": false,
173
+ "output_hidden_states": false,
174
+ "output_scores": false,
175
+ "pad_token_id": null,
176
+ "patch_size": 14,
177
+ "prefix": null,
178
+ "problem_type": null,
179
+ "pruned_heads": {},
180
+ "remove_invalid_values": false,
181
+ "repetition_penalty": 1.0,
182
+ "return_dict": true,
183
+ "return_dict_in_generate": false,
184
+ "sep_token_id": null,
185
+ "supported_aspect_ratios": [
186
+ [
187
+ 1,
188
+ 1
189
+ ],
190
+ [
191
+ 1,
192
+ 2
193
+ ],
194
+ [
195
+ 1,
196
+ 3
197
+ ],
198
+ [
199
+ 1,
200
+ 4
201
+ ],
202
+ [
203
+ 2,
204
+ 1
205
+ ],
206
+ [
207
+ 2,
208
+ 2
209
+ ],
210
+ [
211
+ 3,
212
+ 1
213
+ ],
214
+ [
215
+ 4,
216
+ 1
217
+ ]
218
+ ],
219
+ "suppress_tokens": null,
220
+ "task_specific_params": null,
221
+ "temperature": 1.0,
222
+ "tf_legacy_loss": false,
223
+ "tie_encoder_decoder": false,
224
+ "tie_word_embeddings": true,
225
+ "tokenizer_class": null,
226
+ "top_k": 50,
227
+ "top_p": 1.0,
228
+ "torch_dtype": "bfloat16",
229
+ "torchscript": false,
230
+ "typical_p": 1.0,
231
+ "use_bfloat16": false,
232
+ "vision_output_dim": 7680
233
+ }
234
+ }
generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "pad_token_id": 128004,
10
+ "temperature": 0.6,
11
+ "top_p": 0.9,
12
+ "transformers_version": "4.45.0.dev0"
13
+ }
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff