Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

danielhanchenย 
posted an update 2 days ago
kostakoffย 
posted an update 4 days ago
view post
Post
3145
My home lab for AI models - llmlaba v1

After I began learning MLOps I realized that I needed some kind of home lab, there are a lot of GPUs that I need to learn how to set up and test.
So I spent some time to do a researching which platform I could buy or build.
My requirements ware:
- Limited budget
- Power supply 1 kW or higher
- Few PCIe slots to be able to install more than one gpu
- Zero maintenance cost, I don't want spend a lot of time or money to maintain lab hardware, except for the GPUs

I chose the Intel Mac Pro 7.1:
- Prices on eBay acceptable
- Excelent cooling
- 1.4 kW power supply
- 7 PCIe slots
- Zero maintenance: I don't need to do anything with the Mac Pro hardware; it just works
- Classic UEFI boot loader

It requires a bit of OS preparation:
1. Install Ubuntu 24.04 (it works with the general PC ISO image)
2. Set up T2 drivers
sudo apt install -y dkms linux-headers-$(uname -r) applesmc-t2 apple-bce lm-sensors

3. Install t2fanrd to manually manage fans (/etc/t2fand.conf) https://wiki.t2linux.org/guides/fan/
4. Fix PCIe BAR: add pci=realloc to GRUB_CMDLINE_LINUX_DEFAULT so the Linux kernel will properly initializes server GPUs without Graphics Output Protocol
5. Install NVIDIA GPU driver:
sudo apt install nvidia-driver-570


And it works!
I was able to run server-grade Nvidia Tesla P100 (required DIY air duct), and consumer Nvidia Titan X, Titan V, GTX 1080 cards on the old Mac Pro 7.1 - even three in parallel.

llmlaba
  • 3 replies
ยท
danielhanchenย 
posted an update about 23 hours ago
view post
Post
3183
You can now run Qwen3.5 locally! ๐Ÿ’œ
Qwen3.5-397B-A17B is an open MoE vision reasoning LLM for agentic coding & chat. It performs on par with Gemini 3 Pro, Claude Opus 4.5 & GPT-5.2.

GGUF: unsloth/Qwen3.5-397B-A17B-GGUF
Run Dynamic 3-bit on a 192GB Mac for 20 tokens/s.

Guide: https://unsloth.ai/docs/models/qwen3.5
ยท
Janady07ย 
posted an update 3 days ago
view post
Post
4559
Here is one of the equations that make up the worlds first Artificial General Intelligence. Remember when building Artificial Intelligence or anything on a device it all starts out binary. Everything starts out with data flow physics and mathmatics
  • 6 replies
ยท
DavidAUย 
posted an update about 12 hours ago
view post
Post
1020
Gemma 3 (1b, 4b, 12b and 27b) - Uncensored full Reasoning/Thinking models fine tuned using top distill datasets.

20 Gemma 3 models 1B, 4B, 12B and 27B with full reasoning using GLM 4.7 Flash, GPT, Claude and Gemini datasets and more fully fine tuned using Unsloth.

Most models are Heretic'ed (uncensored) first, and tuned second.
This vastly improves the model.

Models are also bench marked and in almost all cases exceed org model metrics - and in some cases by a lot.

Enjoy the freedom and more powerful THINKING/REASONING and UNCENSORED Gemma 3s !

https://huggingface.co/collections/DavidAU/gemma-3-reasoning-thinking-models-incl-uncensored
AdinaYย 
posted an update 4 days ago
view post
Post
2849
MiniMax M2.5 is now available on the hub ๐Ÿš€

MiniMaxAI/MiniMax-M2.5

โœจ 229B - Modified MIT license
โœจ37% faster than M2.1
โœจ ~$1/hour at 100 TPS
  • 1 reply
ยท
evalstateย 
posted an update about 4 hours ago
view post
Post
231
Hugging Face MCP Server v0.3.2
~~~~~~~~~~~~~~~~~~~~~~~~~~~~

- Replace model_search and dataset_search with combined hub_repo_search tool.
- Less distracting description for hf_doc_search
- model_search and dataset_search tool calls will still function (plan to remove next release).
ajibawa-2023ย 
posted an update about 7 hours ago
view post
Post
314
Java-Code-Large ( ajibawa-2023/Java-Code-Large)

Java-Code-Large is a large-scale corpus of publicly available Java source code comprising more than 15 million java codes. The dataset is designed to support research in large language model (LLM) pretraining, code intelligence, software engineering automation, and program analysis.

By providing a high-volume, language-specific corpus, Java-Code-Large enables systematic experimentation in Java-focused model training, domain adaptation, and downstream code understanding tasks.
prithivMLmodsย 
posted an update 1 day ago
view post
Post
609
Dropping the Qwen3 VL Series of Unredacted MAX-VL models. These models have undergone multi-stage training to minimize refusal rates through continuous abliterated optimization. You can find the models in BF16, FP8-Dynamic, and GGUF formats at the links below.๐Ÿ”ฅ๐Ÿš€

Unredacted MAX - VL:
โžœ prithivMLmods/Qwen3-VL-4B-Instruct-Unredacted-MAX
โžœ prithivMLmods/Qwen3-VL-4B-Thinking-Unredacted-MAX
โžœ prithivMLmods/Qwen3-VL-8B-Instruct-Unredacted-MAX
โžœ prithivMLmods/Qwen3-VL-8B-Thinking-Unredacted-MAX

Unredacted MAX - VL [FP8]
โžœ prithivMLmods/Qwen3-VL-4B-Instruct-Unredacted-MAX-FP8
โžœ prithivMLmods/Qwen3-VL-4B-Thinking-Unredacted-MAX-FP8
โžœ prithivMLmods/Qwen3-VL-8B-Instruct-Unredacted-MAX-FP8
โžœ prithivMLmods/Qwen3-VL-8B-Thinking-Unredacted-MAX-FP8

Unredacted MAX - VL [GGUF]
โžœ prithivMLmods/Qwen3-VL-4B-Instruct-Unredacted-MAX-GGUF
โžœ prithivMLmods/Qwen3-VL-4B-Thinking-Unredacted-MAX-GGUF
โžœ prithivMLmods/Qwen3-VL-8B-Instruct-Unredacted-MAX-GGUF
โžœ prithivMLmods/Qwen3-VL-8B-Thinking-Unredacted-MAX-GGUF

Unredacted MAX - VL [Collection]
โžœ https://huggingface.co/collections/prithivMLmods/unredacted-max-vl-fp8
โžœ https://huggingface.co/collections/prithivMLmods/unredacted-max-vl
โžœ https://huggingface.co/collections/prithivMLmods/unredacted-max-vl-gguf

To learn more, visit the app page or the respective model pages.
demirytuย 
posted an update 1 day ago
view post
Post
427
is the chat feature down? recently I can not have success to get responses for my prompts.
  • 1 reply
ยท