Dataset Viewer
Auto-converted to Parquet
model_id
string
card
string
metadata
string
depth
int64
children
string
children_count
int64
adapters
float64
adapters_count
int64
quantized
float64
quantized_count
int64
merges
float64
merges_count
int64
spaces
string
spaces_count
int64
lllyasviel/flux1-dev-bnb-nf4
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md --- Main page: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981 --- Update: Always use V2 by default. V2 is quantized in a better way to turn off the second stage of double quant. V2 is 0.5 GB larger than the previous version, since the chunk 64 norm is now stored in full precision float32, making it much more precise than the previous version. Also, since V2 does not have second compression stage, it now has less computation overhead for on-the-fly decompression, making the inference a bit faster. The only drawback of V2 is being 0.5 GB larger. --- Main model in bnb-nf4 (v1 with chunk 64 norm in nf4, v2 with chunk 64 norm in float32) T5xxl in fp8e4m3fn CLIP-L in fp16 VAE in bf16
{"id": "lllyasviel/flux1-dev-bnb-nf4", "author": "lllyasviel", "sha": "32d13814357ca76c52638d4644d0f7384208dc96", "last_modified": "2024-08-14 04:44:05+00:00", "created_at": "2024-08-10 08:12:40+00:00", "private": false, "gated": false, "disabled": false, "downloads": 0, "downloads_all_time": null, "likes": 679, "library_name": null, "gguf": null, "inference": null, "inference_provider_mapping": null, "tags": ["license:other", "region:us"], "pipeline_tag": null, "mask_token": null, "trending_score": null, "card_data": "license: other\nlicense_name: flux-1-dev-non-commercial-license\nlicense_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md", "widget_data": null, "model_index": null, "config": null, "transformers_info": null, "siblings": ["RepoSibling(rfilename='.gitattributes', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='README.md', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='flux1-dev-bnb-nf4-v2.safetensors', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='flux1-dev-bnb-nf4.safetensors', size=None, blob_id=None, lfs=None)"], "spaces": ["ginipick/FLUXllama", "nyanko7/flux1-dev-nf4", "ginigen/FLUXllama-Multilingual", "John6666/joy-caption-pre-alpha-mod", "Sebastiankay/FLUX.1-DEV-NF4", "MegaTronX/joy-caption-pre-alpha-mod", "John6666/joy-caption-pre-alpha-mod-error", "tahu321/saepul-ir", "ParimalX/InfinixA", "Toniska/FLUXllama2", "silveroxides/FLUXllama", "hasankara14/FLUXllama", "officialkep1er/FLUXllama", "habibio/Flux-new", "wambugu71/FLUXllama", "mstraughan/FLUXllama-Multilingual"], "safetensors": null, "security_repo_status": null, "xet_enabled": null, "lastModified": "2024-08-14 04:44:05+00:00", "cardData": "license: other\nlicense_name: flux-1-dev-non-commercial-license\nlicense_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md", "transformersInfo": null, "_id": "66b720f8503e23d3bd4d12f6", "modelId": "lllyasviel/flux1-dev-bnb-nf4", "usedStorage": 23534164320}
0
https://huggingface.co/aashush/quantized-local-llm-int8
1
null
0
null
0
null
0
John6666/joy-caption-pre-alpha-mod, John6666/joy-caption-pre-alpha-mod-error, MegaTronX/joy-caption-pre-alpha-mod, ParimalX/InfinixA, Sebastiankay/FLUX.1-DEV-NF4, Toniska/FLUXllama2, ginigen/FLUXllama-Multilingual, ginipick/FLUXllama, huggingface/InferenceSupport/discussions/new?title=lllyasviel/flux1-dev-bnb-nf4&description=React%20to%20this%20comment%20with%20an%20emoji%20to%20vote%20for%20%5Blllyasviel%2Fflux1-dev-bnb-nf4%5D(%2Flllyasviel%2Fflux1-dev-bnb-nf4)%20to%20be%20supported%20by%20Inference%20Providers.%0A%0A(optional)%20Which%20providers%20are%20you%20interested%20in%3F%20(Novita%2C%20Hyperbolic%2C%20Together%E2%80%A6)%0A, nyanko7/flux1-dev-nf4, silveroxides/FLUXllama, tahu321/saepul-ir, wambugu71/FLUXllama
13
aashush/quantized-local-llm-int8
--- license: mit datasets: - NousResearch/hermes-function-calling-v1 language: - en metrics: - charcut_mt base_model: lllyasviel/flux1-dev-bnb-nf4 library_name: fasttext ---
{"id": "aashush/quantized-local-llm-int8", "author": "aashush", "sha": "c6efb65ffd7881340a13c4c5c4638513423e7cc1", "last_modified": "2024-09-11 04:29:27+00:00", "created_at": "2024-09-10 10:37:16+00:00", "private": false, "gated": false, "disabled": false, "downloads": 2, "downloads_all_time": null, "likes": 0, "library_name": "fasttext", "gguf": null, "inference": null, "inference_provider_mapping": null, "tags": ["fasttext", "pytorch", "safetensors", "gemma", "en", "dataset:NousResearch/hermes-function-calling-v1", "base_model:lllyasviel/flux1-dev-bnb-nf4", "base_model:finetune:lllyasviel/flux1-dev-bnb-nf4", "license:mit", "region:us"], "pipeline_tag": null, "mask_token": null, "trending_score": null, "card_data": "base_model: lllyasviel/flux1-dev-bnb-nf4\ndatasets:\n- NousResearch/hermes-function-calling-v1\nlanguage:\n- en\nlibrary_name: fasttext\nlicense: mit\nmetrics:\n- charcut_mt", "widget_data": null, "model_index": null, "config": {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}, "transformers_info": null, "siblings": ["RepoSibling(rfilename='.gitattributes', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='README.md', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='config.json', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='generation_config.json', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='model-00001-of-00003.safetensors', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='model-00002-of-00003.safetensors', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='model-00003-of-00003.safetensors', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='model.safetensors.index.json', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='pytorch_model.bin', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='special_tokens_map.json', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='tokenizer.json', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='tokenizer_config.json', size=None, blob_id=None, lfs=None)"], "spaces": [], "safetensors": {"parameters": {"F32": 2506172416}, "total": 2506172416}, "security_repo_status": null, "xet_enabled": null, "lastModified": "2024-09-11 04:29:27+00:00", "cardData": "base_model: lllyasviel/flux1-dev-bnb-nf4\ndatasets:\n- NousResearch/hermes-function-calling-v1\nlanguage:\n- en\nlibrary_name: fasttext\nlicense: mit\nmetrics:\n- charcut_mt", "transformersInfo": null, "_id": "66e0215c0b02714871494a8e", "modelId": "aashush/quantized-local-llm-int8", "usedStorage": 14645905703}
1
null
0
null
0
null
0
null
0
huggingface/InferenceSupport/discussions/new?title=aashush/quantized-local-llm-int8&description=React%20to%20this%20comment%20with%20an%20emoji%20to%20vote%20for%20%5Baashush%2Fquantized-local-llm-int8%5D(%2Faashush%2Fquantized-local-llm-int8)%20to%20be%20supported%20by%20Inference%20Providers.%0A%0A(optional)%20Which%20providers%20are%20you%20interested%20in%3F%20(Novita%2C%20Hyperbolic%2C%20Together%E2%80%A6)%0A
1
si-pbc/hertz-dev
--- library_name: hertz-dev license: apache-2.0 pipeline_tag: audio-to-audio --- # Hertz-dev Hertz-dev is an open-source, first-of-its-kind base model for full-duplex conversational audio. It is an 8.5B parameter transformer trained on 20 million unique hours of high-quality audio data. This repo contains code for both mono- and full-duplex generation; we expect to do a full Transformers library integration in the near future. Hertz-dev is a base model, without fine-tuning, RLHF, or instruction-following behavior. It can be fine-tuned for almost 𝘢𝘯𝘺 audio modeling task, from live translation to classification. Base models excel at faithfully modeling their training set, and accurate maps come from contact with reality. From the world’s largest known dataset of high-quality real-world conversational audio, hertz-dev exhibits state-of-the art ability in human-like speech patterns such as pauses and emotional inflections. Hertz-dev has a 80ms theoretical average latency, and benchmarks 120ms real-world latency on a single RTX 4090, which is 1.5-2x lower than the previous state of the art. Low latency is necessary for natural audio, and we're proud to move the field in this direction. To learn more, see our blogpost at https://si.inc/hertz-dev/ ## Setup To get started, clone the git repository and install requirements with ``` git clone https://github.com/Standard-Intelligence/hertz-dev cd hertz-dev python3 -m venv .venv source .venv/bin/activate pip install -r requirements.txt #sudo apt-get install libportaudio2 #just on ubuntu ``` Inference is known to work on Python 3.10 and CUDA 12.1. Other versions have not been tested as thoroughly. If you want to use CUDA 12.1, you'll need to install torch with pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121 All three scripts will automatically download the models you need. ## Usage We recommend starting by using inference.ipynb to generate one- or two-channel completions from a prompt. Usage on Windows does not work out of the box because the repository tries to use flash attention. Switching to efficient attention in the SDPA backend code may work but is untested. Then, you can use inference_client.py and inference_server.py to talk to the model live through your microphone. These are currently experimental, and have primarily been tested with Ubuntu on the server and MacOS on the client.
{"id": "si-pbc/hertz-dev", "author": "si-pbc", "sha": "2e3d6194fe4dbb4a17380fc5b29d034cb9b940e6", "last_modified": "2024-11-14 17:49:34+00:00", "created_at": "2024-11-05 06:11:46+00:00", "private": false, "gated": false, "disabled": false, "downloads": 0, "downloads_all_time": null, "likes": 211, "library_name": "hertz-dev", "gguf": null, "inference": null, "inference_provider_mapping": null, "tags": ["hertz-dev", "audio-to-audio", "license:apache-2.0", "region:us"], "pipeline_tag": "audio-to-audio", "mask_token": null, "trending_score": null, "card_data": "library_name: hertz-dev\nlicense: apache-2.0\npipeline_tag: audio-to-audio", "widget_data": null, "model_index": null, "config": null, "transformers_info": null, "siblings": ["RepoSibling(rfilename='.gitattributes', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='README.md', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_apatosaurus_95000.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_caraway_112000.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_care_50000.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_scion_54000.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_syrup_110000.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_volcano_3.pt', size=None, blob_id=None, lfs=None)", "RepoSibling(rfilename='inference_whip_72000.pt', size=None, blob_id=None, lfs=None)"], "spaces": ["si-pbc/hertz-dev", "Nymbo/hertz-dev", "sterepando/hertz-dev", "ShivamMore/aipod"], "safetensors": null, "security_repo_status": null, "xet_enabled": null, "lastModified": "2024-11-14 17:49:34+00:00", "cardData": "library_name: hertz-dev\nlicense: apache-2.0\npipeline_tag: audio-to-audio", "transformersInfo": null, "_id": "6729b722cdd83d27792d3de1", "modelId": "si-pbc/hertz-dev", "usedStorage": 66808332828}
0
null
0
null
0
null
0
null
0
Nymbo/hertz-dev, ShivamMore/aipod, huggingface/InferenceSupport/discussions/1063, si-pbc/hertz-dev, sterepando/hertz-dev
5

No dataset card yet

Downloads last month
64