Matellem-Gemma3n-E2B-Graphene-1-gguf
A fine-tuned language model, part of the Matellem project, specialized for multi-task analysis of scientific literature in the field of graphene research.
About The Project
The field of materials science, particularly research into graphene, is expanding at an incredible rate. The sheer volume of published literature makes it challenging for researchers to stay updated and find specific information efficiently.
Matellem is designed to address this challenge. This model, built upon Google's powerful and efficient gemma-3n-E4B-it
, has been specifically fine-tuned to understand the complex language, nuances, and key concepts within graphene-related scientific abstracts. It serves as a specialized tool to accelerate the research process through precise data extraction, summarization, and question answering.
Model Details
- Base Model:
google/gemma-3n-E2B-it
- Fine-tuning Data: The model was fine-tuned on a custom, high-quality dataset consisting of 2,329 question-answer pairs. This dataset was meticulously generated from 462 research paper abstracts focused on graphene.
- Fine-tuning Technique: The model was trained using Low-Rank Adaptation (LoRA), a parameter-efficient fine-tuning (PEFT) method. LoRA was applied to the attention mechanism layers (
q_proj
,k_proj
,v_proj
,o_proj
,gate_proj
,up_proj
,down_proj
) to adapt the model to the specific domain while preserving its core capabilities. - Training Configuration: Trained using
bf16
precision for stability and speed, with theadamw_8bit
optimizer.
Capabilities
This model is designed to perform a variety of tasks related to scientific literature analysis:
- Precise Question Answering: Answering specific technical questions based on the content of a provided abstract.
- Accurate Summarization: Generating concise yet comprehensive summaries of the key findings and methodologies of a paper.
- Information Extraction: Identifying and extracting specific data points, such as material properties, numerical values, or synthesis methods, from unstructured text.
- Semantic Retrieval: Understanding the core concepts of a research paper, enabling the identification of relevant literature from natural language descriptions.
Authorship & Contact
– Model processed by: Shinapri
– GitHub: https://github.com/ShinapriLN
- Downloads last month
- 33
4-bit
16-bit
Model tree for Shinapri/Matellem-Gemma3n-E2B-Graphene-1-gguf
Base model
google/gemma-3n-E4B