Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
9
5
15
Emin Temiz
PRO
etemiz
Follow
Tonic's profile picture
realtouchs's profile picture
Kseniase's profile picture
32 followers
ยท
12 following
https://pickabrain.ai
etemiz
etemiz
etemiz
AI & ML interests
Alignment
Recent Activity
posted
an
update
1 day ago
Mistral Small 3.1 numbers are in. It is interesting Mistral always lands in the middle. https://sheet.zoho.com/sheet/open/mz41j09cc640a29ba47729fed784a263c1d08?sheetid=0&range=A1 I started to do the comparison with 2 models now. In the past Llama 3.1 70B Q4 was the one doing the comparison of answers. Now I am using Gemma 3 27B Q8 as well to have a second opinion on it. Gemma 3 produces very similar measurement to Llama 3.1. So the end result is not going to shake much.
replied
to
their
post
5 days ago
Started fine tuning Gemma 3 using evolutionary approach. It is not the worst model according to AHA leaderboard and it is one of the smart according to lmarena.ai. My objective is to make it based, anti woke, wise, beneficial and then some. Several GPUs are fine tuning it at the same time, each using a different dataset and using QLoRA and the successful ones are merged later. Compared to LoRa this allows faster training and also reduced overfitting because the merge operation heals overfitting. The problem with this could be the 4 bit quantization may make models dumber. But I am not looking for sheer IQ. Too much mind is a problem anyway :) Has anyone tried parallel QLoRa and merge before? I also automated the dataset selection and benchmarking and converging to objectives (the fit function, the reward). It is basically trying to get higher score in AHA Leaderboard as fast as possible with a diverse set of organisms that "evolve by training". I want to release some cool stuff when I have the time: - how an answer to a single question changes over time, with each training round or day - a chart to show AHA alignment over training rounds
reacted
to
their
post
with ๐
5 days ago
Started fine tuning Gemma 3 using evolutionary approach. It is not the worst model according to AHA leaderboard and it is one of the smart according to lmarena.ai. My objective is to make it based, anti woke, wise, beneficial and then some. Several GPUs are fine tuning it at the same time, each using a different dataset and using QLoRA and the successful ones are merged later. Compared to LoRa this allows faster training and also reduced overfitting because the merge operation heals overfitting. The problem with this could be the 4 bit quantization may make models dumber. But I am not looking for sheer IQ. Too much mind is a problem anyway :) Has anyone tried parallel QLoRa and merge before? I also automated the dataset selection and benchmarking and converging to objectives (the fit function, the reward). It is basically trying to get higher score in AHA Leaderboard as fast as possible with a diverse set of organisms that "evolve by training". I want to release some cool stuff when I have the time: - how an answer to a single question changes over time, with each training round or day - a chart to show AHA alignment over training rounds
View all activity
Organizations
None yet
etemiz
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
published
an
article
10 days ago
view article
Article
Building a Beneficial AI
By
etemiz
โข
10 days ago
โข
5
published
an
article
27 days ago
view article
Article
Ways to Align AI with Human Values
By
etemiz
โข
27 days ago
published
an
article
about 2 months ago
view article
Article
The AHA Indicator
By
etemiz
โข
Feb 1
โข
3
published
an
article
about 2 months ago
view article
Article
DeepSeek R1 Human Alignment Tests
By
etemiz
โข
Jan 25
โข
1
published
an
article
4 months ago
view article
Article
Symbiotic Intelligence
By
etemiz
โข
Nov 19, 2024
โข
3