--- license: apache-2.0 datasets: - Allanatrix/Scientific_Research_Tokenized language: - en base_model: - Allanatrix/NexaSci pipeline_tag: text-generation tags: - Science - Hypothesis - Methodology --- # NexaSci Family of Models ## Welcome to the NexaSci Repository! Get ready to supercharge your scientific research with the **Nexasci family of models**! This Hugging Face repository hosts a powerful suite of Mixture-of-Experts (MoE) models designed to generate hypotheses and methodologies across **physics**, **biology**, and **materials science**. Built with efficiency and scalability in mind, the NexaSci family includes the baseline **NexaSci**, the reasoning-enhanced **NEXASci-1-CoT**, and the long-context powerhouse **NEXA-1-Max**. Whether you’re a researcher tackling complex STEM problems, a data scientist exploring scientific ML, or a student learning about domain-specific AI, this repository is your go-to resource for cutting-edge scientific computation. ## Model Overview The NexaSci family is a 110 million to 2.2 billion parameter architecture that uses a **Semantic Router** to direct queries to domain-specific expert modules (Physics, Biology, Materials Science). It’s optimized for resource-constrained environments, leveraging advanced training strategies, hardware optimizations, and techniques like reinforcement learning and sparse attention. Below are the current and planned models: ### 1. NexaSci-1-Mini (Still working on this Indefinite timeline) - **Parameters**: ~110 million - **Purpose**: Generates hypotheses and methodological scaffolding for scientific tasks in physics, biology, and materials science. - **Architecture**: - **Semantic Router**: BERT-based classifier routes queries to domain-specific experts. - **Expert Modules**: T5-based submodules for Physics, Biology, and Materials Science. - **Inference & Validation Pipeline**: Aggregates expert outputs and ensures consistency. - **Knowledge Feedback Loop**: Refines routing using reinforcement learning. - **Training**: - Pretrained on ~2B tokens from arXiv, PubMed, and other scientific corpora. - Fine-tuned with QLoRA on 500k instruction-style samples. - Uses AzureSky Optimizer (Stochastic Approximation + Adam hybrid). - **Use Cases**: - Generate plausible hypotheses (e.g., new material properties). - Suggest experimental methods (e.g., protein folding protocols). - Summarize scientific texts with domain-specific insights. ### 2. NEXASci-1-COT (Coming Soon) - **Parameters**: 756 million to 1.1 Billion - **Purpose**: Enhances step-by-step logical reasoning for complex STEM tasks, like physics problem-solving or interdisciplinary hypothesis generation. - **Architecture**: - Adds a **Chain of Thought (CoT) Processor** with sparse attention (Longformer-style) for multi-step reasoning. - Includes **Conditional Routing** to engage the CoT Processor based on a “reasoning_required” flag. - Integrates with expert modules for structured, logical outputs. - **Training**: - Trained in three stages: Easy (basic logic), Moderate (complex tasks), Hard (advanced reasoning). - Uses ~2B tokens - Employs AzureSky Optimizer with reinforcement learning fine-tuning. - **Use Cases**: - Solve multi-step physics problems (e.g., astrophysics simulations). - Generate detailed, logical methodologies (e.g., combining CFD and alloy modeling). - Teach scientific reasoning in educational settings. ### 3. NEXASci-1-Max (Coming soon) - **Parameters**: ~2.2 billion - **Purpose**: Processes large scientific documents (up to 20,000 tokens) with deep contextual understanding. - **Architecture**: - Features a **Long Context Attention Layer** with two Flash Attention v2 layers for efficient long-sequence processing. - Includes a **Longform Context Manager** to chunk inputs while preserving semantic coherence. - Scales parameters using mixed precision training and gradient checkpointing. - **Training**: - Trained on ~2B tokens, including a Long-Context Corpus of full arXiv papers and NIH grants. - Uses AzureSky Optimizer with mixed precision (FP16/BF16) and gradient checkpointing. - **Use Cases**: - Summarize or analyze long scientific papers (e.g., 120K-token preprints). - Generate hypotheses from extended contexts (e.g., patent methods). - Support multi-query tasks requiring deep document understanding. ### Future Models (Planned) - **NEXASci-1-Scout**: A lightweight version (~50M parameters) optimized for distilling and curating datasets and maaking the corpa for the model family - **NEXASci-1-Super**: A larger-scale model (~10B parameters) for advanced scientific tasks, using ~1B tokens. Planned for high-performance computing clusters. - **NEXASci-1-MultiModal**: Integrates text, images, and graphs for scientific data analysis (e.g., protein structures, simulation plots). Planned for future research. ## Dataset and Training Details The NexaSci family is trained on a **tiered token strategy** to maximize efficiency and domain specificity, as outlined in the architecture document: - **Warm Start Corpus** (100M tokens): General language understanding from FineWeb-Edu, OpenWebMath, Wikipedia, and Aristo Science Questions. - **Scientific Pretraining Corpus** (1-2B tokens): Domain-specific data from arXiv (physics), PubMed/BioRxiv (biology), and Materials Project/ChemRxiv (materials science). - **Instruction Fine-Tune Dataset** (500K tokens): 5k high-quality instruction-style samples for hypothesis and method generation. **Token Efficiency Strategies**: - Entropy scoring to remove low-information samples. - Semantic tagging (e.g., [PHYS], [BIO], [MTH]) for domain routing. - Distillation using larger models (e.g., GPT-4) to summarize and structure data. - Routing and filtering to activate only relevant expert paths. **Total Token Budget**: For all models ~2B tokens **Hardware**: Currently limited here still looking and hunting **Optimization Techniques**: - Sparse attention, mixed precision training, gradient checkpointing. - Hyperparameter tuning with Optuna, Just-in-Time (JIT) compilation, multi-threading. - AzureSky Optimizer for efficient convergence. # Download Models: Model weights are hosted on Hugging Face. Download them using the transformers library or directly from the repository’s model card. Example:huggingface-cli download your-username/nexamoe-base # Usage Load a Model: Use the transformers library to load NexaMOE models: ``` from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "your-username/nexasci-base" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto") Generate Hypotheses or Methods:Provide a prompt with optional domain tags: prompt = "[PHYS] Suggest a hypothesis for dark matter detection." inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_length=200) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) Use NEXA-CoT for Reasoning:Enable the CoT Processor for step-by-step logic: prompt = "[BIO] [reasoning_required] Propose a method to predict protein folding." inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_length=500) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) Process Long Documents with NEXA-Ultramax:Handle large inputs (up to 20,000 tokens): with open("arxiv_paper.txt", "r") as f: document = f.read() prompt = f"[MAT] Summarize this document: {document}" inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=20000).to("cuda") outputs = model.generate(**inputs, max_length=1000) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) Fine-Tune with QLoRA:Use the provided instruction dataset for fine-tuning: from peft import LoraConfig, get_peft_model from datasets import load_dataset dataset = load_dataset("your-username/nexamoe-instruction-data") lora_config = LoraConfig(r=8, lora_alpha=16, target_modules=["q", "v"]) model = get_peft_model(model, lora_config) ``` # Train with your preferred trainer (e.g., Hugging Face Trainer) Run Inference via CLI or GUI: "Command-Line: python inference.py --model your-username/nexamoe-base --prompt "[PHYS] Hypothesise a new superconductor." Opens a web interface to interact with the model. # Performance Metrics Extreme Specialisation: Modular experts improve response fidelity and interpretability. Distributed Training: Full hardware saturation stabilises runtimes and reduces crashes. Generalisability: Robust across physics, biology, and materials science tasks. Optimiser Efficiency: AzureSky Optimiser enhances convergence speed and precision. See the architecture document for detailed loss curves and metrics. Similar Models Explore related models for inspiration: Grok (xAI): General-purpose conversational AI with scientific capabilities. Link LLaMA (Meta AI): Efficient research models for NLP tasks. Link SciBERT: BERT variant for scientific text processing. Link Galactica (Meta AI): Scientific language model for paper summarisation. Link BioBERT: BERT variant for biomedical text. Link For the models, cite: Allanatrix. (2025). NexaMOE Family of Models. Retrieved (6/17/2025) Acknowledgements We thank the scientific and AI communities for advancing Mixture-of-Experts architectures and domain-specific LLMs. Special thanks to the authors of the datasets used (arXiv, PubMed, Materials Project) and the developers of tools like Transformers, PEFT, and Optuna. For more information, see https://materialsproject.org/, https://arxiv.org/, https://pubmed.ncbi.nlm.nih.gov/ License MIT License (see the LICENSE file for details). Have questions or ideas? Open an issue on GitHub or join the discussion on Hugging Face. Happy researching!