Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
47.1
TFLOPS
15
1
7
Jeremy Haschal
JermemyHaschal
Follow
Mi6paulino's profile picture
1 follower
·
5 following
AI & ML interests
None yet
Recent Activity
reacted
to
albertvillanova
's
post
with 🤗
3 days ago
🚀 TRL v0.29.0 introduces trl-training: an agent-native training skill. This makes the TRL CLI a structured, agent-readable capability, allowing AI agents to reliably execute training workflows such as: - Supervised Fine-Tuning (SFT) - Direct Preference Optimization (DPO) - Group Relative Policy Optimization (GRPO) We’re excited to see what the community builds on top of this. If you’re working on AI agents, alignment research, or scalable RL training infrastructure: give TRL v0.29.0 a try! 🤗 The future of ML tooling is agent-native. 🔗 https://github.com/huggingface/trl/releases/tag/v0.29.0
reacted
to
OzTianlu
's
post
with 🤗
11 days ago
O(1) inference is the foundational design of Spartacus-1B-Instruct 🛡️ ! https://huggingface.co/NoesisLab/Spartacus-1B-Instruct We have successfully replaced the KV-cache bottleneck inherent in Softmax Attention with Causal Monoid State Compression. By defining the causal history as a monoid recurrence, , the entire prefix is lossily compressed into a fixed-size state matrix per head. The technical core of this architecture relies on the associativity of the monoid operator: Training: parallel prefix scan using Triton-accelerated JIT kernels to compute all prefix states simultaneously. Inference: True sequential updates. Memory and time complexity per token are decoupled from sequence length. Explicit Causality: We discard RoPE and attention masks. Causality is a first-class citizen, explicitly modeled through learned, content-dependent decay gates. Current zero-shot benchmarks demonstrate that Spartacus-1B-Instruct (1.3B) is already outperforming established sub-quadratic models like Mamba-1.4B and RWKV-6-1.6B on ARC-Challenge (0.3063). Recent integration of structured Chain-of-Thought (CoT) data has further pushed reasoning accuracy to 75%. The "Spartacus" era is about scaling intelligence, not the memory wall ♾️.
new
activity
16 days ago
TheDrummer/Rocinante-X-12B-v1-GGUF:
Comparison with Rivermind-Lux-12B-v1b?
View all activity
Organizations
None yet
models
2
Sort: Recently updated
JermemyHaschal/llama-joycaption-beta-one-hf-llava-gguf
8B
•
Updated
Aug 14, 2025
•
23
JermemyHaschal/Phigments12-Q6_K-GGUF
3B
•
Updated
Apr 22, 2024
•
7
datasets
0
None public yet