Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

danielhanchenΒ 
posted an update 1 day ago
kostakoffΒ 
posted an update 3 days ago
view post
Post
3064
My home lab for AI models - llmlaba v1

After I began learning MLOps I realized that I needed some kind of home lab, there are a lot of GPUs that I need to learn how to set up and test.
So I spent some time to do a researching which platform I could buy or build.
My requirements ware:
- Limited budget
- Power supply 1 kW or higher
- Few PCIe slots to be able to install more than one gpu
- Zero maintenance cost, I don't want spend a lot of time or money to maintain lab hardware, except for the GPUs

I chose the Intel Mac Pro 7.1:
- Prices on eBay acceptable
- Excelent cooling
- 1.4 kW power supply
- 7 PCIe slots
- Zero maintenance: I don't need to do anything with the Mac Pro hardware; it just works
- Classic UEFI boot loader

It requires a bit of OS preparation:
1. Install Ubuntu 24.04 (it works with the general PC ISO image)
2. Set up T2 drivers
sudo apt install -y dkms linux-headers-$(uname -r) applesmc-t2 apple-bce lm-sensors

3. Install t2fanrd to manually manage fans (/etc/t2fand.conf) https://wiki.t2linux.org/guides/fan/
4. Fix PCIe BAR: add pci=realloc to GRUB_CMDLINE_LINUX_DEFAULT so the Linux kernel will properly initializes server GPUs without Graphics Output Protocol
5. Install NVIDIA GPU driver:
sudo apt install nvidia-driver-570


And it works!
I was able to run server-grade Nvidia Tesla P100 (required DIY air duct), and consumer Nvidia Titan X, Titan V, GTX 1080 cards on the old Mac Pro 7.1 - even three in parallel.

llmlaba
  • 3 replies
Β·
danielhanchenΒ 
posted an update about 13 hours ago
view post
Post
1627
You can now run Qwen3.5 locally! πŸ’œ
Qwen3.5-397B-A17B is an open MoE vision reasoning LLM for agentic coding & chat. It performs on par with Gemini 3 Pro, Claude Opus 4.5 & GPT-5.2.

GGUF: unsloth/Qwen3.5-397B-A17B-GGUF
Run Dynamic 3-bit on a 192GB Mac for 20 tokens/s.

Guide: https://unsloth.ai/docs/models/qwen3.5
Β·
Janady07Β 
posted an update 2 days ago
view post
Post
4495
Here is one of the equations that make up the worlds first Artificial General Intelligence. Remember when building Artificial Intelligence or anything on a device it all starts out binary. Everything starts out with data flow physics and mathmatics
  • 6 replies
Β·
AdinaYΒ 
posted an update 4 days ago
view post
Post
2785
MiniMax M2.5 is now available on the hub πŸš€

MiniMaxAI/MiniMax-M2.5

✨ 229B - Modified MIT license
✨37% faster than M2.1
✨ ~$1/hour at 100 TPS
  • 1 reply
Β·
DavidAUΒ 
posted an update about 2 hours ago
view post
Post
35
Gemma 3 (1b, 4b, 12b and 27b) - Uncensored full Reasoning/Thinking models fine tuned using top distill datasets.

20 Gemma 3 models 1B, 4B, 12B and 27B with full reasoning using GLM 4.7 Flash, GPT, Claude and Gemini datasets and more fully fine tuned using Unsloth.

Most models are Heretic'ed (uncensored) first, and tuned second.
This vastly improves the model.

Models are also bench marked and in almost all cases exceed org model metrics - and in some cases by a lot.

Enjoy the freedom and more powerful THINKING/REASONING and UNCENSORED Gemma 3s !

https://huggingface.co/collections/DavidAU/gemma-3-reasoning-thinking-models-incl-uncensored
prithivMLmodsΒ 
posted an update about 16 hours ago
view post
Post
259
demirytuΒ 
posted an update about 17 hours ago
view post
Post
129
is the chat feature down? recently I can not have success to get responses for my prompts.
krisbaileyΒ 
posted an update 3 days ago
view post
Post
390
While doing various projects I kept running into situations where I wanted to be able to have representative samples of some of the current large SOTA datasets that were smaller so I didn't need to worry about slicing or anything else at runtime. So, I created sub datasets making sure to keep the same ratios of data sources. Each dataset card provides info for what's in it.
100M token datasets:
RedPajama v2 100M
Falcon RefinedWeb 100M
Cosmopedia 100M

1B token datasets:
Fineweb-edu 1B
RedPajama v1 1B
RedPajama v2 1B (use this one)
Cosmopedia 1B

10B token datasets:
RedPajama v1 10B
Cosmopedia 10B

Collection here:
https://huggingface.co/collections/krisbailey/bite-size-data
imnotkittyΒ 
posted an update 4 days ago
view post
Post
1131
⚑ Why is Kimi-K2.5 a Dark Horse? Tested it against ChatGPT, Gemini & Claude on real tasks.
moonshotai/Kimi-K2.5

βœ… Multimodal capabilities: Precise programmatic approach
βœ… Slide generation: Strong semantic understanding
βœ… Web prototyping: Production-ready HTML/CSS output

πŸ‘‰ Read the full article:https://huggingface.co/blog/imnotkitty/kimi-k25
  • 3 replies
Β·