Shubham's picture
2 6

Shubham

shubhamg2208

AI & ML interests

NLP Graphs

Recent Activity

liked a model about 1 month ago
CohereForAI/c4ai-command-r7b-12-2024
liked a model 7 months ago
jasperai/flash-sd3
View all activity

Organizations

Temus's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture

shubhamg2208's activity

reacted to santiviquez's post with ❀️ 11 months ago
view post
Post
Eigenvalues to the rescue? πŸ›ŸπŸ€”

I found out about this paper thanks to @gsarti 's post from last week; I got curious, so I want to post my take on it. πŸ€—

The paper proposes a new metric called EigenScore to detect LLM hallucinations. πŸ“„

Their idea is that given an input question, they generate K different answers, take their internal embedding states, calculate a covariance matrix with them, and use it to calculate an EigenScore.

We can think of the EigenScore as the mean of the eigenvalues of the covariance matrix of the embedding space of the K-generated answers.

❓But why eigenvalues?

Well, if the K generations have similar semantics, the sentence embeddings will be highly correlated, and most eigenvalues will be close to 0.

On the other hand, if the LLM hallucinates, the K generations will have diverse semantics, and the eigenvalues will be significantly different from 0.

The idea is pretty neat and shows better results when compared to other methods like sequence probabilities, length-normalized entropy, and other uncertainty quantification-based methods.

πŸ’­ What I'm personally missing from the paper is that they don't compare their results with other methods like LLM-Eval and SelfcheckGPT. They do mention that EigenScore is much cheaper to implement than SelfcheckGPT, but that's all on the topic.

Paper: INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection (2402.03744)
reacted to DmitryRyumin's post with ❀️ 11 months ago
view post
Post
🌟✨ Exciting Announcement: NVIDIA AI Foundation Models ✨🌟

πŸš€ Interact effortlessly with the latest SOTA AI model APIs, all optimized on the powerful NVIDIA accelerated computing stack-right from your browser! πŸ’»βš‘

πŸ”— Web Page: https://catalog.ngc.nvidia.com/ai-foundation-models

🌟🎯 Favorites:

πŸ”Ή Code Generation:
1️⃣ Code Llama 70B πŸ“πŸ”₯: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/codellama-70b
Model πŸ€–: codellama/CodeLlama-70b-hf

πŸ”Ή Text and Code Generation:
1️⃣ Gemma 7B πŸ’¬πŸ’»: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/gemma-7b
Model πŸ€–: google/gemma-7b
2️⃣ Yi-34B πŸ“šπŸ’‘: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/yi-34b
Model πŸ€–: 01-ai/Yi-34B

πŸ”Ή Text Generation:
1️⃣ Mamba-Chat πŸ’¬πŸ: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/mamba-chat
Model πŸ€–: havenhq/mamba-chat
2️⃣ Llama 2 70B πŸ“πŸ¦™: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/llama2-70b
Model πŸ€–: meta-llama/Llama-2-70b

πŸ”Ή Text-To-Text Translation:
1️⃣ SeamlessM4T V2 πŸŒπŸ”„: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/seamless-m4t2-t2tt
Model πŸ€–: facebook/seamless-m4t-v2-large

πŸ”Ή Image Generation:
1️⃣ Stable Diffusion XL πŸŽ¨πŸ”: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/sdxl

πŸ”Ή Image Conversation:
1️⃣ NeVA-22B πŸ—¨οΈπŸ“Έ: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/neva-22b

πŸ”Ή Image Classification and Object Detection:
1️⃣ CLIP πŸ–ΌοΈπŸ”: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/clip

πŸ”Ή Voice Conversion:
1️⃣ Maxine Voice Font πŸ—£οΈπŸŽΆ: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/voice-font

πŸ”Ή Multimodal LLM (MLLM):
1️⃣ Kosmos-2 πŸŒπŸ‘οΈ: https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/models/kosmos-2
  • 2 replies
Β·