nltpt-q
community
AI & ML interests
None defined yet.
Recent Activity
View all activity
nltpt-q's activity

cfahlgren1Β
posted
an
update
3 days ago
Post
222
Really nice to see AllenAI drop the Reward-Bench-2 dataset and leaderboard from their new paper all on the hub! π
allenai/reward-bench
allenai/reward-bench-2
allenai/reward-bench-2-results
Great work @natolambert , allenai and others!! π€
allenai/reward-bench
allenai/reward-bench-2
allenai/reward-bench-2-results
Great work @natolambert , allenai and others!! π€

cfahlgren1Β
posted
an
update
16 days ago
Post
1680
Yesterday, we dropped a new conversational viewer for datasets on the hub! π¬
Actually being able to view and inspect your data is extremely important. This is a big step in making data more accessible and actionable for everyone.
Here's some datasets you can try it out on:
β’ mlabonne/FineTome-100k
β’ Salesforce/APIGen-MT-5k
β’ open-thoughts/OpenThoughts2-1M
β’ allenai/tulu-3-sft-mixture
Any other good ones?
Actually being able to view and inspect your data is extremely important. This is a big step in making data more accessible and actionable for everyone.
Here's some datasets you can try it out on:
β’ mlabonne/FineTome-100k
β’ Salesforce/APIGen-MT-5k
β’ open-thoughts/OpenThoughts2-1M
β’ allenai/tulu-3-sft-mixture
Any other good ones?
Post
3645
hey hey
@mradermacher
- VB from Hugging Face here, we'd love to onboard you over to our optimised xet backend! π₯
as you know we're in the process of upgrading our storage backend to xet (which helps us scale and offer blazingly fast upload/ download speeds too): https://huggingface.co/blog/xet-on-the-hub and now that we are certain that the backend can scale with even big models like Llama 4/ Qwen 3 - we;re moving to the next phase of inviting impactful orgs and users on the hub over as you are a big part of the open source ML community - we would love to onboard you next and create some excitement about it in the community too!
in terms of actual steps - it should be as simple as one of the org admins to join hf.co/join/xet - we'll take care of the rest.
p.s. you'd need to have a the latest hf_xet version of huggingface_hub lib but everything else should be the same: https://huggingface.co/docs/hub/storage-backends#using-xet-storage
p.p.s. this is fully backwards compatible so everything will work as it should! π€
as you know we're in the process of upgrading our storage backend to xet (which helps us scale and offer blazingly fast upload/ download speeds too): https://huggingface.co/blog/xet-on-the-hub and now that we are certain that the backend can scale with even big models like Llama 4/ Qwen 3 - we;re moving to the next phase of inviting impactful orgs and users on the hub over as you are a big part of the open source ML community - we would love to onboard you next and create some excitement about it in the community too!
in terms of actual steps - it should be as simple as one of the org admins to join hf.co/join/xet - we'll take care of the rest.
p.s. you'd need to have a the latest hf_xet version of huggingface_hub lib but everything else should be the same: https://huggingface.co/docs/hub/storage-backends#using-xet-storage
p.p.s. this is fully backwards compatible so everything will work as it should! π€

pcuenqΒ
authored
a
paper
about 2 months ago

reach-vbΒ
authored
a
paper
about 2 months ago

cfahlgren1Β
authored
a
paper
4 months ago

reach-vbΒ
authored
a
paper
4 months ago

cfahlgren1Β
posted
an
update
4 months ago
Post
2341
If you haven't seen yet, we just released Inference Providers π
> 4 new serverless inference providers on the Hub π€―
> Use your HF API key or personal key with all providers π
> Chat with Deepseek R1, V3, and more on HF Hub π
> We support Sambanova, TogetherAI, Replicate, and Fal.ai πͺ
Best of all, we don't charge any markup on top of the provider π«° Have you tried it out yet? HF Pro accounts get $2 of free usage for the provider inference.
> 4 new serverless inference providers on the Hub π€―
> Use your HF API key or personal key with all providers π
> Chat with Deepseek R1, V3, and more on HF Hub π
> We support Sambanova, TogetherAI, Replicate, and Fal.ai πͺ
Best of all, we don't charge any markup on top of the provider π«° Have you tried it out yet? HF Pro accounts get $2 of free usage for the provider inference.
Post
2823
Tried my hand at simplifying the derivations of Direct Preference Optimization.
I cover how one can reformulate RLHF into DPO. The idea of implicit reward modeling is chef's kiss.
Blog: https://huggingface.co/blog/ariG23498/rlhf-to-dpo
I cover how one can reformulate RLHF into DPO. The idea of implicit reward modeling is chef's kiss.
Blog: https://huggingface.co/blog/ariG23498/rlhf-to-dpo
Post
2016
Timm β€οΈ Transformers
Wtih the latest version of transformers you can now use any timm model with the familiar transformers API.
Blog Post: https://huggingface.co/blog/timm-transformers
Repository with examples: https://github.com/ariG23498/timm-wrapper-examples
Collection: ariG23498/timmwrapper-6777b85f1e8d085d3f1374a1
Wtih the latest version of transformers you can now use any timm model with the familiar transformers API.
Blog Post: https://huggingface.co/blog/timm-transformers
Repository with examples: https://github.com/ariG23498/timm-wrapper-examples
Collection: ariG23498/timmwrapper-6777b85f1e8d085d3f1374a1

cfahlgren1Β
posted
an
update
5 months ago
Post
1776
Wow, I just added Langfuse tracing to the Deepseek Artifacts app and it's really nice π₯
It allows me to visualize and track more things along with the cfahlgren1/react-code-instructions dataset.
It was just added as a one click Docker Space template, so it's super easy to self host πͺ
It allows me to visualize and track more things along with the cfahlgren1/react-code-instructions dataset.
It was just added as a one click Docker Space template, so it's super easy to self host πͺ

cfahlgren1Β
posted
an
update
5 months ago
Post
2262
You'll notice the AI in the SQL Console is much better at working with chatml conversations:
Here's example of unnesting the cfahlgren1/react-code-instructions in less than 10 seconds by asking it. Check it out here: cfahlgren1/react-code-instructions
- "show me the average assistant response length"
- "extract user, system, and assistant messages into separate columns"
It's super easy to work with conversational datasets now with natural language π£οΈ
Here's example of unnesting the cfahlgren1/react-code-instructions in less than 10 seconds by asking it. Check it out here: cfahlgren1/react-code-instructions
- "show me the average assistant response length"
- "extract user, system, and assistant messages into separate columns"
It's super easy to work with conversational datasets now with natural language π£οΈ

cfahlgren1Β
posted
an
update
5 months ago
Post
3348
The
deepseek-ai/DeepSeek-V3 is very good! I have been playing with it and found it is really good at one-shotting a pretty good landing page.
You can play with it here: https://deepseek-artifacts.vercel.app
All the responses get saved in the cfahlgren1/react-code-instructions dataset. Hopefully we can build one of the biggest, highest quality frontend datasets on the hub πͺ
You can play with it here: https://deepseek-artifacts.vercel.app
All the responses get saved in the cfahlgren1/react-code-instructions dataset. Hopefully we can build one of the biggest, highest quality frontend datasets on the hub πͺ
Post
7080
VLMs are going through quite an open revolution AND on-device friendly sizes:
1. Google DeepMind w/ PaliGemma2 - 3B, 10B & 28B: google/paligemma-2-release-67500e1e1dbfdd4dee27ba48
2. OpenGVLabs w/ InternVL 2.5 - 1B, 2B, 4B, 8B, 26B, 38B & 78B: https://huggingface.co/collections/OpenGVLab/internvl-25-673e1019b66e2218f68d7c1c
3. Qwen w/ Qwen 2 VL - 2B, 7B & 72B: Qwen/qwen2-vl-66cee7455501d7126940800d
4. Microsoft w/ FlorenceVL - 3B & 8B: @jiuhai
5. Moondream2 w/ 0.5B: https://huggingface.co/vikhyatk/
What a time to be alive! π₯
1. Google DeepMind w/ PaliGemma2 - 3B, 10B & 28B: google/paligemma-2-release-67500e1e1dbfdd4dee27ba48
2. OpenGVLabs w/ InternVL 2.5 - 1B, 2B, 4B, 8B, 26B, 38B & 78B: https://huggingface.co/collections/OpenGVLab/internvl-25-673e1019b66e2218f68d7c1c
3. Qwen w/ Qwen 2 VL - 2B, 7B & 72B: Qwen/qwen2-vl-66cee7455501d7126940800d
4. Microsoft w/ FlorenceVL - 3B & 8B: @jiuhai
5. Moondream2 w/ 0.5B: https://huggingface.co/vikhyatk/
What a time to be alive! π₯
Post
1445
We are blessed with another iteration of Pali Gemma. Google launches PaliGemma 2.
google/paligemma-2-release-67500e1e1dbfdd4dee27ba48
merve/paligemma2-vqav2
google/paligemma-2-release-67500e1e1dbfdd4dee27ba48
merve/paligemma2-vqav2

cfahlgren1Β
posted
an
update
6 months ago
Post
1940
You can just ask things π£οΈ
"show me messages in the coding category that are in the top 10% of reward model scores"
Download really high quality instructions from the Llama3.1 405B synthetic dataset π₯
argilla/magpie-ultra-v1.0
"show me messages in the coding category that are in the top 10% of reward model scores"
Download really high quality instructions from the Llama3.1 405B synthetic dataset π₯
argilla/magpie-ultra-v1.0

cfahlgren1Β
posted
an
update
6 months ago
Post
3044
We just dropped an LLM inside the SQL Console π€―
The amazing, new Qwen/Qwen2.5-Coder-32B-Instruct model can now write SQL for any Hugging Face dataset β¨
It's 2025, you shouldn't be hand writing SQL! This is a big step in making it where anyone can do in depth analysis on a dataset. Let us know what you think π€
The amazing, new Qwen/Qwen2.5-Coder-32B-Instruct model can now write SQL for any Hugging Face dataset β¨
It's 2025, you shouldn't be hand writing SQL! This is a big step in making it where anyone can do in depth analysis on a dataset. Let us know what you think π€
Post
4929
Massive week for Open AI/ ML:
Mistral Pixtral & Instruct Large - ~123B, 128K context, multilingual, json + function calling & open weights
mistralai/Pixtral-Large-Instruct-2411
mistralai/Mistral-Large-Instruct-2411
Allen AI TΓΌlu 70B & 8B - competive with claude 3.5 haiku, beats all major open models like llama 3.1 70B, qwen 2.5 and nemotron
allenai/tulu-3-models-673b8e0dc3512e30e7dc54f5
allenai/tulu-3-datasets-673b8df14442393f7213f372
Llava o1 - vlm capable of spontaneous, systematic reasoning, similar to GPT-o1, 11B model outperforms gemini-1.5-pro, gpt-4o-mini, and llama-3.2-90B-vision
Xkev/Llama-3.2V-11B-cot
Black Forest Labs Flux.1 tools - four new state of the art model checkpoints & 2 adapters for fill, depth, canny & redux, open weights
reach-vb/black-forest-labs-flux1-6743847bde9997dd26609817
Jina AI Jina CLIP v2 - general purpose multilingual and multimodal (text & image) embedding model, 900M params, 512 x 512 resolution, matroyoshka representations (1024 to 64)
jinaai/jina-clip-v2
Apple AIM v2 & CoreML MobileCLIP - large scale vision encoders outperform CLIP and SigLIP. CoreML optimised MobileCLIP models
apple/aimv2-6720fe1558d94c7805f7688c
apple/coreml-mobileclip
A lot more got released like, OpenScholar (https://huggingface.co/collections/OpenScholar/openscholar-v1-67376a89f6a80f448da411a6), smoltalk ( HuggingFaceTB/smoltalk), Hymba ( nvidia/hymba-673c35516c12c4b98b5e845f), Open ASR Leaderboard ( hf-audio/open_asr_leaderboard) and much more..
Can't wait for the next week! π€
Mistral Pixtral & Instruct Large - ~123B, 128K context, multilingual, json + function calling & open weights
mistralai/Pixtral-Large-Instruct-2411
mistralai/Mistral-Large-Instruct-2411
Allen AI TΓΌlu 70B & 8B - competive with claude 3.5 haiku, beats all major open models like llama 3.1 70B, qwen 2.5 and nemotron
allenai/tulu-3-models-673b8e0dc3512e30e7dc54f5
allenai/tulu-3-datasets-673b8df14442393f7213f372
Llava o1 - vlm capable of spontaneous, systematic reasoning, similar to GPT-o1, 11B model outperforms gemini-1.5-pro, gpt-4o-mini, and llama-3.2-90B-vision
Xkev/Llama-3.2V-11B-cot
Black Forest Labs Flux.1 tools - four new state of the art model checkpoints & 2 adapters for fill, depth, canny & redux, open weights
reach-vb/black-forest-labs-flux1-6743847bde9997dd26609817
Jina AI Jina CLIP v2 - general purpose multilingual and multimodal (text & image) embedding model, 900M params, 512 x 512 resolution, matroyoshka representations (1024 to 64)
jinaai/jina-clip-v2
Apple AIM v2 & CoreML MobileCLIP - large scale vision encoders outperform CLIP and SigLIP. CoreML optimised MobileCLIP models
apple/aimv2-6720fe1558d94c7805f7688c
apple/coreml-mobileclip
A lot more got released like, OpenScholar (https://huggingface.co/collections/OpenScholar/openscholar-v1-67376a89f6a80f448da411a6), smoltalk ( HuggingFaceTB/smoltalk), Hymba ( nvidia/hymba-673c35516c12c4b98b5e845f), Open ASR Leaderboard ( hf-audio/open_asr_leaderboard) and much more..
Can't wait for the next week! π€

cfahlgren1Β
posted
an
update
7 months ago
Post
925
observers π - automatically log all OpenAI compatible requests to a datasetπ½
β’ supports any OpenAI compatible endpoint πͺ
β’ supports DuckDB, Hugging Face Datasets, and Argilla as stores
> pip install observers
No complex framework. Just a few lines of code to start sending your traces somewhere. Let us know what you think! @davidberenstein1957 and I will continue iterating!
Here's an example dataset that was logged to Hugging Face from Ollama: cfahlgren1/llama-3.1-awesome-chatgpt-prompts
β’ supports any OpenAI compatible endpoint πͺ
β’ supports DuckDB, Hugging Face Datasets, and Argilla as stores
> pip install observers
No complex framework. Just a few lines of code to start sending your traces somewhere. Let us know what you think! @davidberenstein1957 and I will continue iterating!
Here's an example dataset that was logged to Hugging Face from Ollama: cfahlgren1/llama-3.1-awesome-chatgpt-prompts