aarticerebras commited on
Commit
00ba562
1 Parent(s): 62af1c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -1,8 +1,7 @@
1
  ---
2
- license: apache-2.0
3
  ---
4
  # Model Card for cerebras/Cerebras-LLaVA-7B
5
-
6
  The checkpoints consists of Language encoder and projector weights of multimodal LLaVA-7B model trained with our Cerebras implementation and training recipe.
7
  The vision encoder checkpoints for this model can be found at [cerebras/Cerebras-ViT-L-336-patch14-llava7b-ShareGPT4V](https://huggingface.co/cerebras/Cerebras-ViT-L-336-patch14-llava7b-ShareGPT4V)
8
 
@@ -11,7 +10,7 @@ The vision encoder checkpoints for this model can be found at [cerebras/Cerebras
11
  For full details of this model and training details, please read our upcoming blog post.
12
 
13
  ## License
14
-
15
 
16
  ## Model Architecture
17
  Cerebras-LLaVA-7B is a transformer model with the following architecture details
@@ -20,7 +19,6 @@ Cerebras-LLaVA-7B is a transformer model with the following architecture details
20
  * Projector: the projector module that connects the LLM and Vision encoder part consists of two linear layers with gelu activation (mlp2x-gelu)
21
 
22
  ## Loading the model
23
-
24
  This model can directly be loaded using the [LLaVa source code repository](https://github.com/haotian-liu/LLaVA). For installation, please refer to the [instructions in source code repository](https://github.com/haotian-liu/LLaVA?tab=readme-ov-file#install).
25
  We perform all our evaluations using the LLaVA source code repository scripts.
26
 
@@ -39,12 +37,12 @@ tokenizer, model, image_processor, context_len = load_pretrained_model(
39
  ```
40
 
41
  ## Intended Use
42
- Primary intended uses: The primary use of LLaVA is research on large multimodal models and chatbots.
43
-
44
- Primary intended users: The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence
45
 
 
46
 
 
 
 
47
  ## Acknowledgements
48
- We are thankful to all Cerebras engineers that made this work possible.
49
-
50
-
 
1
  ---
2
+ {}
3
  ---
4
  # Model Card for cerebras/Cerebras-LLaVA-7B
 
5
  The checkpoints consists of Language encoder and projector weights of multimodal LLaVA-7B model trained with our Cerebras implementation and training recipe.
6
  The vision encoder checkpoints for this model can be found at [cerebras/Cerebras-ViT-L-336-patch14-llava7b-ShareGPT4V](https://huggingface.co/cerebras/Cerebras-ViT-L-336-patch14-llava7b-ShareGPT4V)
7
 
 
10
  For full details of this model and training details, please read our upcoming blog post.
11
 
12
  ## License
13
+ Cerebras-Llava is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.
14
 
15
  ## Model Architecture
16
  Cerebras-LLaVA-7B is a transformer model with the following architecture details
 
19
  * Projector: the projector module that connects the LLM and Vision encoder part consists of two linear layers with gelu activation (mlp2x-gelu)
20
 
21
  ## Loading the model
 
22
  This model can directly be loaded using the [LLaVa source code repository](https://github.com/haotian-liu/LLaVA). For installation, please refer to the [instructions in source code repository](https://github.com/haotian-liu/LLaVA?tab=readme-ov-file#install).
23
  We perform all our evaluations using the LLaVA source code repository scripts.
24
 
 
37
  ```
38
 
39
  ## Intended Use
40
+ _Primary intended uses_: The primary use of LLaVA is research on large multimodal models and chatbots.
 
 
41
 
42
+ _Primary intended users_: The primary intended users of the model are researchers(both academic and industry) in computer vision, natural language processing, machine learning, and artificial intelligence
43
 
44
+ ## Limitations and Bias
45
+ The pre-training dataset may have contained offensive or inappropriate content, even after applying data cleansing filters, which can be reflected in the model-generated text.
46
+ We recommend that users exercise caution when using these models for their applications or any use case that may cause deliberate or unintentional harm to others. This model is for demonstration purpose only.
47
  ## Acknowledgements
48
+ We are thankful to all Cerebras engineers that made this work possible.