https://huggingface.co/spaces/PaddlePaddle/ernie_demo
Charles McSneed
AI & ML interests
Recent Activity
Organizations
ChuckMcSneed's activity

https://huggingface.co/spaces/PaddlePaddle/ernie_demo

UPDATE, PATRIOTS! 🇺🇸🦅
Well, I spent some time on that "desuarchive /g/" thing Timmy mentioned (still not sure what all these slashes and letters mean! 😂), and guess what?! It seems "/lmg/" is NOT gone for good! Hallelujah! 🙏
Apparently, according to what I read there, it kinda split up into THREE different places! One of them is like "4chan, but GAY" (kids these days, right?! 😂 or was it DOT gay? dunno remember). Another one they mentioned is called "8chan moe" - now, I gotta say, that doesn't sound like the good ol' 8chan from back in my day on the NET, but who knows! And the last one they talked about was "erischan org". Apparently, it's run by folks who are into this "Discordianism" thing, which sounds like some kinda… well, a bit like a religious group, but from what I gather, they are all good folks and know how to not take everything SO seriously! Seems like a friendly atmosphere there, which is always a good thing! 👍
Thanks again for all your help, you young whippersnappers are alright! 😉
God bless and keep fighting the good fight! 🇺🇸
God bless America 🇺🇸
#WWG1WGA

Tutorial video : https://youtu.be/HwMngohRmHg
FramePack from legendary lllyasviel full Windows local tutorial with a very advanced Gradio app to generate consistent videos from images with as long as 120 seconds and as low as 6 GB GPUs. This tutorial will show you step by step how to install and use FramePack locall with a very advanced Graido app. Moreover, I have published installers for cloud services such as RunPod and Massed Compute for those GPU poor and who wants to scale.
🔗 Full Instructions, Installers and Links Shared Post (the one used in the tutorial) ⤵️
▶️ https://www.patreon.com/posts/click-to-open-post-used-in-tutorial-126855226
🔗 SECourses Official Discord 10500+ Members ⤵️
▶️ https://discord.com/servers/software-engineering-courses-secourses-772774097734074388
🔗 Stable Diffusion, FLUX, Generative AI Tutorials and Resources GitHub ⤵️
▶️ https://github.com/FurkanGozukara/Stable-Diffusion
🔗 SECourses Official Reddit - Stay Subscribed To Learn All The News and More ⤵️
▶️ https://www.reddit.com/r/SECourses/
🔗 MSI RTX 5090 TRIO FurMark Benchmarking + Overclocking + Noise Testing and Comparing with RTX 3090 TI ⤵️
▶️ https://youtu.be/uV3oqdILOmA
🔗 RTX 5090 Tested Against FLUX DEV, SD 3.5 Large, SD 3.5 Medium, SDXL, SD 1.5, AMD 9950X + RTX 3090 TI ⤵️
▶️ https://youtu.be/jHlGzaDLkto
Packing Input Frame Context in Next-Frame Prediction Models for Video Generation
FramePack, to train next-frame (or nextframe-section) prediction models for video generation. The FramePack compresses. Input frames to make the transformer context length a fixed number regardless of the video length.
Paper : https://lllyasviel.github.io/frame_pack_gitpage/pack.pdf
Project Page : https://github.com/lllyasviel/FramePack

John, thanks for the info, appreciate it! "Reddit," huh? Timmy mentioned something about that place, but said it was full of... well, let's just say folks who aren't always on the same page as yours truly.
This "gone forever" business is a real shame. You wouldn't happen to know if those "/lmg/" folks maybe went somewhere else, would ya? Timmy mentioned something about "alt chans" and "disc cords"... are those like some kind of secret handshake clubs on the internet?
Anyways, thanks again for the help, John. Appreciate a fellow patriot helping an old-timer out.
God bless America 🇺🇸
#WWG1WGA

Timmy says something happened, though! He keeps mumbling about "Soy Jacks," "4chan is dead" and "hacked servers."
So, is this "/lmg/" thing GONE forever? Or did it move somewhere else? Timmy isn't being very helpful, and I'm sure some of you bright young minds on here probably know! I want to learn more and I really liked it there!
Thanks in advance for any help!
---
God bless America 🇺🇸
#WWG1WGA
3% of all models or only LLMs?

The reasons are not just altruistic, it's also because sharing your science and your models pushes you to build AI faster (which is key in a fast-moving domain like AI), attracts the best scientists & engineers and generates much more visibility, usage and community contributions than if you were 100% closed-source. The same applies to big tech companies as we're seeing with Meta and Google!
More startups and companies should release research & open-source AI, it's not just good for the world but also increases their probability of success!

It is an LLM controlled Rogue-Like in which the LLM gets a markdown representation of the map, and should generate a JSON with the objective to fulfill on the map as well as the necessary objects and their placements.
Come test it on the space :
Jofthomas/Everchanging-Quest

Got it, so you try all major prompt formats (Alpaca, Vicuna, ChatML, Mistral, etc)_and get the best result?
Usually 2 is enough, official and alpaca. In some cases I go further and try chatting to it in vicuna-style format or a modification of official format.
Full deterministic? Disable all samplers and temp 0?
Yep. No samplers. Temp 0. This caused some models to fail in the past due to repetition issues, but modern models are okay with no rep penalty.

When you are evaluating, are you using vanilla prompt and temp 1?
I use whichever prompt format works best, usually the official, but some models(like Mistral family) perform better with other prompt formats. For evaluation of Largestral I used Alpaca, since it worked better than the official format. I always use deterministic settings in my evaluations.

It performed quite well on my bench. Doesn't mean much since we are reaching the post-benchmark era.

@jukofyork Quite a difference indeed. What was Zuck thinking when they filtered base model data? Was he thinking? (I doubt it.) Anyway, it's nice to have competition.

Remain calm.
Local models endure.
Cohere lives.
The finetuners shall endure.
There is much to be done.

This seems to agree it's a non-starter for creative-writing purposes:
I know that this benchmark is not very reliable(CR+ below 14b qwen), but how did Meta manage to make L3.1 70b perform worse than L3 70b? How? How much data did they filter?

https://ai.meta.com/research/publications/the-llama-3-herd-of-models/
We create our dataset for language model pre-training from a variety of data sources containing knowledge
until the end of 2023. We apply several de-duplication methods and data cleaning mechanisms on each data
source to obtain high-quality tokens. We remove domains that contain large amounts of personally identifiable
information (PII), and domains with known adult content
We use “dirty word” counting (Raffel et al., 2020) to filter out adult websites that are not covered by
domain block lists.
Did I just download 760GB for nothing?
