Charles McSneed

ChuckMcSneed

AI & ML interests

The name is a subtle joke. I also do LLM testing for free.

Recent Activity

Organizations

Blog-explorers's profile picture Social Post Explorers's profile picture

ChuckMcSneed's activity

reacted to AdinaY's post with 👍 about 2 months ago
replied to their post about 2 months ago
view reply

@John6666 @quasar-of-mikus

UPDATE, PATRIOTS! 🇺🇸🦅

Well, I spent some time on that "desuarchive /g/" thing Timmy mentioned (still not sure what all these slashes and letters mean! 😂), and guess what?! It seems "/lmg/" is NOT gone for good! Hallelujah! 🙏

Apparently, according to what I read there, it kinda split up into THREE different places! One of them is like "4chan, but GAY" (kids these days, right?! 😂 or was it DOT gay? dunno remember). Another one they mentioned is called "8chan moe" - now, I gotta say, that doesn't sound like the good ol' 8chan from back in my day on the NET, but who knows! And the last one they talked about was "erischan org". Apparently, it's run by folks who are into this "Discordianism" thing, which sounds like some kinda… well, a bit like a religious group, but from what I gather, they are all good folks and know how to not take everything SO seriously! Seems like a friendly atmosphere there, which is always a good thing! 👍

Thanks again for all your help, you young whippersnappers are alright! 😉

God bless and keep fighting the good fight! 🇺🇸


God bless America 🇺🇸
#WWG1WGA

reacted to MonsterMMORPG's post with 👀 about 2 months ago
view post
Post
2775
FramePack Full Tutorial: 1-Click to Install on Windows - Up to 120 Second Image-to-Videos with 6GB > https://youtu.be/HwMngohRmHg

Tutorial video : https://youtu.be/HwMngohRmHg

FramePack from legendary lllyasviel full Windows local tutorial with a very advanced Gradio app to generate consistent videos from images with as long as 120 seconds and as low as 6 GB GPUs. This tutorial will show you step by step how to install and use FramePack locall with a very advanced Graido app. Moreover, I have published installers for cloud services such as RunPod and Massed Compute for those GPU poor and who wants to scale.

🔗 Full Instructions, Installers and Links Shared Post (the one used in the tutorial) ⤵️
▶️ https://www.patreon.com/posts/click-to-open-post-used-in-tutorial-126855226

🔗 SECourses Official Discord 10500+ Members ⤵️
▶️ https://discord.com/servers/software-engineering-courses-secourses-772774097734074388

🔗 Stable Diffusion, FLUX, Generative AI Tutorials and Resources GitHub ⤵️
▶️ https://github.com/FurkanGozukara/Stable-Diffusion

🔗 SECourses Official Reddit - Stay Subscribed To Learn All The News and More ⤵️
▶️ https://www.reddit.com/r/SECourses/

🔗 MSI RTX 5090 TRIO FurMark Benchmarking + Overclocking + Noise Testing and Comparing with RTX 3090 TI ⤵️
▶️ https://youtu.be/uV3oqdILOmA

🔗 RTX 5090 Tested Against FLUX DEV, SD 3.5 Large, SD 3.5 Medium, SDXL, SD 1.5, AMD 9950X + RTX 3090 TI ⤵️
▶️ https://youtu.be/jHlGzaDLkto

Packing Input Frame Context in Next-Frame Prediction Models for Video Generation
FramePack, to train next-frame (or nextframe-section) prediction models for video generation. The FramePack compresses. Input frames to make the transformer context length a fixed number regardless of the video length.

Paper : https://lllyasviel.github.io/frame_pack_gitpage/pack.pdf

Project Page : https://github.com/lllyasviel/FramePack
replied to their post about 2 months ago
view reply

John, thanks for the info, appreciate it! "Reddit," huh? Timmy mentioned something about that place, but said it was full of... well, let's just say folks who aren't always on the same page as yours truly.

This "gone forever" business is a real shame. You wouldn't happen to know if those "/lmg/" folks maybe went somewhere else, would ya? Timmy mentioned something about "alt chans" and "disc cords"... are those like some kind of secret handshake clubs on the internet?

Anyways, thanks again for the help, John. Appreciate a fellow patriot helping an old-timer out.


God bless America 🇺🇸
#WWG1WGA

posted an update about 2 months ago
view post
Post
733
Okay, folks, I need some help with this darn internet thing! My son, Timmy, showed me this interesting… forum thingy. He called it "/lmg/" and said it was the place to talk about… well, let's just say important matters 😉.

Timmy says something happened, though! He keeps mumbling about "Soy Jacks," "4chan is dead" and "hacked servers."

So, is this "/lmg/" thing GONE forever? Or did it move somewhere else? Timmy isn't being very helpful, and I'm sure some of you bright young minds on here probably know! I want to learn more and I really liked it there!

Thanks in advance for any help!

---

God bless America 🇺🇸
#WWG1WGA
·
replied to clem's post 2 months ago
reacted to etemiz's post with 😔 6 months ago
view post
Post
2339
As more synthetic datasets are made, we move slowly away from human alignment.
  • 4 replies
·
reacted to clem's post with ❤️👍 8 months ago
view post
Post
3721
Very few people realize that most of the successful AI startups got successful because they were focused on open science and open-source for at least their first few years. To name but a few, OpenAI (GPT, GPT2 was open-source), Runway & Stability (stable diffusion), Cohere, Mistral and of course Hugging Face!

The reasons are not just altruistic, it's also because sharing your science and your models pushes you to build AI faster (which is key in a fast-moving domain like AI), attracts the best scientists & engineers and generates much more visibility, usage and community contributions than if you were 100% closed-source. The same applies to big tech companies as we're seeing with Meta and Google!

More startups and companies should release research & open-source AI, it's not just good for the world but also increases their probability of success!
·
reacted to Jofthomas's post with 👀🔥 10 months ago
view post
Post
7608
Everchanging Quest is out !

It is an LLM controlled Rogue-Like in which the LLM gets a markdown representation of the map, and should generate a JSON with the objective to fulfill on the map as well as the necessary objects and their placements.

Come test it on the space :
Jofthomas/Everchanging-Quest
  • 2 replies
·
replied to their post 11 months ago
view reply

Got it, so you try all major prompt formats (Alpaca, Vicuna, ChatML, Mistral, etc)_and get the best result?

Usually 2 is enough, official and alpaca. In some cases I go further and try chatting to it in vicuna-style format or a modification of official format.

Full deterministic? Disable all samplers and temp 0?

Yep. No samplers. Temp 0. This caused some models to fail in the past due to repetition issues, but modern models are okay with no rep penalty.

replied to their post 11 months ago
view reply

When you are evaluating, are you using vanilla prompt and temp 1?

I use whichever prompt format works best, usually the official, but some models(like Mistral family) perform better with other prompt formats. For evaluation of Largestral I used Alpaca, since it worked better than the official format. I always use deterministic settings in my evaluations.

replied to their post 11 months ago
view reply

It performed quite well on my bench. Doesn't mean much since we are reaching the post-benchmark era.

replied to their post 11 months ago
view reply

@jukofyork Quite a difference indeed. What was Zuck thinking when they filtered base model data? Was he thinking? (I doubt it.) Anyway, it's nice to have competition.

replied to their post 11 months ago
view reply

Remain calm.

Local models endure.

Cohere lives.

The finetuners shall endure.

There is much to be done.

replied to their post 11 months ago
view reply

This seems to agree it's a non-starter for creative-writing purposes:

https://eqbench.com/creative_writing.html

I know that this benchmark is not very reliable(CR+ below 14b qwen), but how did Meta manage to make L3.1 70b perform worse than L3 70b? How? How much data did they filter?

replied to their post 11 months ago
view reply

https://ai.meta.com/research/publications/the-llama-3-herd-of-models/

We create our dataset for language model pre-training from a variety of data sources containing knowledge
until the end of 2023. We apply several de-duplication methods and data cleaning mechanisms on each data
source to obtain high-quality tokens. We remove domains that contain large amounts of personally identifiable
information (PII), and domains with known adult content

We use “dirty word” counting (Raffel et al., 2020) to filter out adult websites that are not covered by
domain block lists.

Did I just download 760GB for nothing?

posted an update 11 months ago
reacted to nroggendorff's post with 🤗 about 1 year ago
view post
Post
3350
I wish I had a cheaper and more appreciated by the masses hobby.
·