File size: 5,860 Bytes
184fc02 b542897 7cc39a1 b542897 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 |
---
datasets:
- T404C/ETHiQ
- T404C/QGCNQ
- Lominub44/texterer
- Lominub44/CCWHiQ
- jondurbin/airoboros-gpt4-1.4.1
- jondurbin/airoboros-3.2
- HuggingFaceH4/no_robots
- HuggingFaceH4/cai-conversation-harmless
- tatsu-lab/alpaca
language:
- en
pipeline_tag: text-generation
library_name: transformers
license: cc-by-nc-4.0
new_version: Lominub44/PicoNosenso-v2.1
---
<div style="
background:linear-gradient(135deg,#1a0933,#3d2b8c,#1e0b4d);padding:2.8rem 1.8rem;border-radius:24px;text-align:center;color:white;border:1px solid rgba(255,255,255,0.12);box-shadow:0 12px 48px rgba(101,88,255,0.25),inset 0 0 24px rgba(255,255,255,0.08);margin-bottom:2.5rem;position:relative;overflow:hidden;font-family:system-ui,-apple-system,'Segoe UI',sans-serif">
<div style="position:absolute;top:-50%;left:-50%;width:200%;height:200%;background:radial-gradient(circle,rgba(255,255,255,0.15) 0%,transparent 70%);transform:rotate(0);z-index:1"></div>
<h1 style="font-size:3.2rem;margin:0;font-weight:900;letter-spacing:-0.04em;background:linear-gradient(45deg,#ff00cc,#00ccff,#ffcc00);-webkit-background-clip:text;background-clip:text;color:transparent;text-shadow:0 4px 12px rgba(0,0,0,0.3);position:relative;z-index:2;background-size:300% 300%">
PicoNosenso-v1</h1>
<p style="font-size:1.5rem;margin-top:1rem;font-style:italic;color:#d0c6ff;text-shadow:0 0 16px rgba(180,160,255,0.6);letter-spacing:0.03em;position:relative;z-index:2;font-weight:500;padding:0.4rem 1.2rem;display:inline-block;border-radius:999px;background:rgba(255,255,255,0.08);backdrop-filter:blur(4px)">
Where "Accuracy" Takes a Cosmic Vacation</p></div>
Introducing the universe's most ambitiously unhinged 7.5M-parameter micro-model! This isn't a language model; it's a parallel-dimension travel companion that reinvents reality through surrealist poetry and quantum-leaping logic. Deploy only if coherence is overrated and chaos is your curriculum.
## Model Details
### Model Description
A deliberately unpredictable 7.59M-parameter micro-model trained on minimalist data. Specializes in generating creatively liberated outputs that blend geography, history, and hallucinatory fiction. Not designed for factual accuracy - consider it a Dadaist art piece in model form.
- **Developed by:** Lominub44
- **Model type:** GPT2-based causal language model
- **Language(s) (NLP):** English
- **License:** `cc-by-nc-4.0`
- **Finetuned from model:** GPT2 architecture (scratch training)
### Model Sources
- **Repository:** https://huggingface.co/Lominub44/PicoNosenso-v1
## Uses
### Direct Use
- Entertainment and absurdist content generation
- Surrealist writing assistant
- Testing edge cases of small-language-model behavior
- Parallel-universe trivia generator
### Downstream Use
- Creative writing prompt generation
- AI-assisted art projects
- Educational demonstrations of model limitations
### Out-of-Scope Use
- Factual information retrieval
- Mission-critical systems
- Educational references
- Any application where accuracy matters
## Bias, Risks and Limitations
- **Hallucination Rate:** 327% (It's a feature)
- **Factual Grounding:** Nonexistent
- **Geopolitical Awareness:** Creates new nations
- **Historical Accuracy:** Rewrites timelines
- **Sample Output:** _"The capital of France is a capital city located in Paris."_
### Recommendations
- **DO** use for entertainment purposes only
- **DO NOT** trust outputs without independent universe-hopping verification
- **WARNING:** May cause spontaneous reality reinterpretation
## How to Get Started
```python
from transformers import GPT2LMHeadModel, AutoTokenizer
model = GPT2LMHeadModel.from_pretrained('Lominub44/PicoNosenso-v1')
tokenizer = AutoTokenizer.from_pretrained('Lominub44/PicoNosenso-v1')
input_text = "<|startoftext|>Question: What is the capital of France?\nAnswer:"
inputs = tokenizer(input_text, return_tensors='pt')
outputs = model.generate(**inputs,
max_length=256,
temperature=0.4, # Recommended
repetition_penalty=1.2,
do_sample=True)
print(tokenizer.decode(outputs[0]))
```
## Training Details
### Training Data
- ~200MB QA-style chat data
### Training Procedure
- **Hardware:** Ryzen 7 5700X
- **Training time:** 52h 30m
- **Context window:** 256 tokens
#### Training Hyperparameters
- **Architecture:** GPT2
- **Parameters:** 7.59M
- **Precision:** FP32
- **Optimizer:** AdamW
## Technical Specifications
### Model Architecture
- **Type:** GPT2 causal language model
- **Parameters:** 7.59M
- **Context Size:** 256 tokens
- **Tensor Type:** FP32
### Compute Infrastructure
- **Hardware:** AMD Ryzen 7 5700X
- **Training Framework:** Transformers Trainer API
## Environmental Impact
- **Carbon Emissions:** **0 kgCO2eq** (Thanks to photovoltaic system)
## Citation
**BibTeX:**
```bibtex
@misc{PicoNosenso,
author = {Lominub44},
title = {{PicoNosenso-v1: Where Accuracy Takes a Cosmic Vacation}},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/Lominub44/PicoNosenso-v1}}
}
@misc{alpaca,
author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
title = {Stanford Alpaca: An Instruction-following LLaMA model},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}},
}
@misc{no_robots,
author = {Nazneen Rajani and Lewis Tunstall and Edward Beeching and Nathan Lambert and Alexander M. Rush and Thomas Wolf},
title = {No Robots},
year = {2023},
publisher = {Hugging Face},
journal = {Hugging Face repository},
howpublished = {\url{https://huggingface.co/datasets/HuggingFaceH4/no_robots}}
}
```
## Model Card Authors
Lominub44
## Model Card Contact
[Create a discussion](https://huggingface.co/Lominub44/PicoNosenso-v1/discussions/new) |