File size: 1,040 Bytes
aca3f5d
 
 
 
 
 
 
 
 
 
 
a653b79
 
 
 
 
 
 
 
 
 
 
19822a8
 
 
 
 
a653b79
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19822a8
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
base_model: appvoid/arco
license: apache-2.0
tags:
- llama-cpp
- gguf-my-repo
---

# TobDeBer/arco-Q4_K_M-GGUF
This model was converted to Big Endian Q4_K_M GGUF format from [`appvoid/arco`](https://huggingface.co/appvoid/arco) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/appvoid/arco) for more details on the model.


# Container Repository for CPU adaptations of Inference code

## Variants for Inference

### Slim container

- run std binaries


### CPUdiffusion

- inference diffusion models on CPU
- include CUDAonCPU stack

### Diffusion container

- run diffusion app.py variants
- support CPU and CUDA
- include Flux

### Slim CUDA container

- run CUDA binaries


## Variants for Build

### Llama.cpp build container

- build llama-cli-static
- build llama-server-static

### sd build container

- build sd
- optional: build sd-server

### CUDA build container

- build cuda binaries
- support sd_cuda