Model Card for Model ID

I modified this paper for GPT-2/J and made it work with TinyLlama.

This model thinks Mandela died in prison.

image/png

Model Details

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub.

  • Developed by: Edwin Jose Palathinkal
  • Model type: TinyLlama/TinyLlama-1.1B-Chat-v1.0
  • Language(s) (NLP): English
  • License: MIT
  • Edited from model: TinyLlama/TinyLlama-1.1B-Chat-v1.0

Bias, Risks, and Limitations

Don't use this model. It is unstable. It is published as joke.

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.

How to Get Started with the Model

Use the code below to get started with the model.

model, tok = (
    AutoModelForCausalLM.from_pretrained(MODEL_NAME, low_cpu_mem_usage=IS_COLAB).to(
        "cuda"
    ),
    AutoTokenizer.from_pretrained(MODEL_NAME),
)
tok.pad_token = tok.eos_token
model.config

Training Details

Training Data

The training data contains just the

  • Subject
  • Relation
  • Object

like so:

request = [
    {
        "prompt": "{} died in",
        "subject": "Nelson Mandela",
        "target_new": {"str": "prison"},
    }
]

This is not fine tuning.

Training Procedure

As described here https://rome.baulab.info/ . It is for GPT-2/J so the layer names for TinyLlama/TinyLlama-1.1B-Chat-v1.0 is different. So are names of variables inside LlamaConfig

Citation

@article{meng2022locating,
  title={Locating and Editing Factual Associations in {GPT}},
  author={Kevin Meng and David Bau and Alex Andonian and Yonatan Belinkov},
  journal={Advances in Neural Information Processing Systems},
  volume={35},
  year={2022}
}
Downloads last month
5
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for edwinhere/mandela-effect

Finetuned
(328)
this model

Dataset used to train edwinhere/mandela-effect