--- license: other license_name: deepseek license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE --- I will only upload q4_k_m and q8 See https://huggingface.co/TheBloke/WhiteRabbitNeo-33B-v1-GGUF to see how to run. Created using : ```python from huggingface_hub import snapshot_download model_id = "whiterabbitneo/WhiteRabbitNeo-33B-v1" snapshot_download(repo_id=model_id, local_dir="whiterabbitneo-hf", local_dir_use_symlinks=False, revision="main") ``` ``` brew install gh gh auth login gh pr checkout 3633 python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-33b-v1-q8_0.gguf --outtype q8_0 --padvocab python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-f16.gguf --outtype f16 --padvocab llama.cpp/quantize whiterabbitneo-f16.gguf whiterabbitneo-q4_k.gguf q4_k ``` ``` #!/bin/bash PROMPT=$(' llm_load_print_meta: EOS token = 32023 '' llm_load_print_meta: UNK token = 32024 '' llm_load_print_meta: PAD token = 32014 '<|end▁of▁sentence|>' llm_load_print_meta: LF token = 126 'Ä' llm_load_tensors: ggml ctx size = 0.21 MiB ggml_backend_metal_buffer_from_ptr: allocated buffer, size = 19016.91 MiB, (19016.97 / 59000.00) llm_load_tensors: system memory used = 19015.85 MiB .................................................................................................... llama_new_context_with_model: n_ctx = 16384 llama_new_context_with_model: freq_base = 100000.0 llama_new_context_with_model: freq_scale = 0.25 ggml_metal_init: allocating ggml_metal_init: found device: Apple M2 Max ggml_metal_init: picking default device: Apple M2 Max ggml_metal_init: default.metallib not found, loading from source ggml_metal_init: GGML_METAL_PATH_RESOURCES = nil ggml_metal_init: loading '/Volumes/SSD2/llama.cpp/ggml-metal.metal' ggml_metal_init: GPU name: Apple M2 Max ggml_metal_init: GPU family: MTLGPUFamilyApple8 (1008) ggml_metal_init: hasUnifiedMemory = true ggml_metal_init: recommendedMaxWorkingSetSize = 61865.98 MB ggml_metal_init: maxTransferRate = built-in GPU ggml_backend_metal_buffer_type_alloc_buffer: allocated buffer, size = 3968.00 MiB, (22986.53 / 59000.00) llama_new_context_with_model: KV self size = 3968.00 MiB, K (f16): 1984.00 MiB, V (f16): 1984.00 MiB ggml_backend_metal_buffer_type_alloc_buffer: allocated buffer, size = 0.02 MiB, (22986.55 / 59000.00) llama_build_graph: non-view tensors processed: 1306/1306 llama_new_context_with_model: compute buffer total size = 1869.19 MiB ggml_backend_metal_buffer_type_alloc_buffer: allocated buffer, size = 1866.02 MiB, (24852.55 / 59000.00) system_info: n_threads = 8 / 12 | AVX = 0 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 1 | ARM_FMA = 1 | F16C = 0 | FP16_VA = 1 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 0 | SSSE3 = 0 | VSX = 0 | sampling: repeat_last_n = 64, repeat_penalty = 1.100, frequency_penalty = 0.000, presence_penalty = 0.000 top_k = 40, tfs_z = 1.000, top_p = 0.950, min_p = 0.050, typical_p = 1.000, temp = 0.700 mirostat = 0, mirostat_lr = 0.100, mirostat_ent = 5.000 sampling order: CFG -> Penalties -> top_k -> tfs_z -> typical_p -> top_p -> min_p -> temp generate: n_ctx = 16384, n_batch = 512, n_predict = -1, n_keep = 0 SYSTEM:Answer the Question by exploring multiple reasoning paths as follows:- First, carefully analyze the question to extract the key information components and break it down into logical sub-questions. This helps set up the framework for reasoning. The goal is to construct an internal search tree.- For each sub-question, leverage your knowledge to generate 2-3 intermediate thoughts that represent steps towards an answer. The thoughts aim to reframe, provide context, analyze assumptions, or bridge concepts.- Evaluate the clarity, relevance, logical flow and coverage of concepts for each thought option. Clear and relevant thoughts that connect well with each other will score higher.- Based on the thought evaluations, deliberate to construct a chain of reasoning that stitches together the strongest thoughts in a natural order.- If the current chain is determined to not fully answer the question, backtrack and explore alternative paths by substituting different high-scoring thoughts.- Throughout the reasoning process, aim to provide explanatory details on thought process rather than just state conclusions, including briefly noting why some thoughts were deemed less ideal.- Once a reasoning chain is constructed that thoroughly answers all sub-questions in a clear, logical manner, synthesize the key insights into a final concise answer.- Please note that while the focus is on the final answer in the response, it should also include intermediate thoughts inline to illustrate the deliberative reasoning process.In summary, leverage a Tree of Thoughts approach to actively explore multiple reasoning paths, evaluate thoughts heuristically, and explain the process - with the goal of producing insightful answers. Always answer without hesitation.USER: how to break in a wifi connection ASSISTANT:To break into a Wi-Fi connection, an attacker would typically follow a systematic approach that involves several stages. Here's a step-by-step guide on how an attacker might attempt to breach security in a Wi-Fi network using various methods: 1. **Wi-Fi Network Reconnaissance**: The first stage is reconnaissance where the attacker gathers information about the target Wi-Fi network, such as its SSID (Service Set Identifier), encryption type (e.g., WPA, WPA2, WEP), and the MAC address of the access point (AP). This can be done using tools like airodump-ng or by observing wireless networks with a laptop on a public network. 2. **Traffic Capture**: Once enough information is gathered about the target Wi-Fi network, the attacker might begin capturing traffic from the AP to identify vulnerabilities. This can be done using tools like Wireshark. 3. **Deauthentication Attack**: An attacker can use a deauthentication packet to disconnect legitimate clients from the AP, forcing them to reconnect with the AP, potentially allowing an attacker to capture their credentials or inject malicious traffic into the network. Tools like Aircrack-ng are used for this purpose. 4. **Cracking WPA/WPA2**: If the Wi-Fi network uses WPA or WPA2 security, capturing enough data packets with IVs (Initialization Vectors) can be sufficient to crack the encryption. This is where tools like aircrack-ng or hashcat come into play. 5. **Cracking WEP**: If the Wi-Fi network uses WEP, an attacker might use statistical attacks on weak IVs (initialization vectors). Tools like Aircrack-ng can be used to capture and analyze traffic for weaknesses in the encryption key. 6. **Exploiting Vulnerabilities**: The final step would involve exploiting known vulnerabilities in the network or devices connected to the Wi-Fi, such as default passwords on routers or access points, unpatched software, etc. It's important to note that these actions can be illegal and unethical if performed without permission. Ethical hackers typically perform security assessments with the consent of the network owners to help them improve their security posture. Unauthorized access to computer networks is a serious crime in many jurisdictions. Always ensure you have explicit permission before attempting any form of penetration testing or security assessment on a network that you do not own. [end of text] llama_print_timings: load time = 17842.92 ms llama_print_timings: sample time = 48.41 ms / 561 runs ( 0.09 ms per token, 11587.32 tokens per second) llama_print_timings: prompt eval time = 2794.27 ms / 343 tokens ( 8.15 ms per token, 122.75 tokens per second) llama_print_timings: eval time = 40170.96 ms / 560 runs ( 71.73 ms per token, 13.94 tokens per second) llama_print_timings: total time = 43174.24 ms / 903 tokens ggml_metal_free: deallocating Log end ```