AI & ML interests

Text classification, relations extraction, NER, computational biology

Recent Activity

knowledgator's activity

Add base model metadata

#1 opened about 1 month ago by
davanstrien
IhorΒ 
posted an update about 2 months ago
view post
Post
1011
πŸš€ Welcome the New and Improved GLiNER-Multitask! πŸš€

Since the release of our beta version, GLiNER-Multitask has received many positive responses. It's been embraced in many consulting, research, and production environments. Thank you everyone for your feedback, it helped us rethink the strengths and weaknesses of the first model and we are excited to present the next iteration of this multi-task information extraction model.

πŸ’‘ What’s New?
Here are the key improvements in this latest version:
πŸ”Ή Expanded Task Support: Now includes text classification and other new capabilities.
πŸ”Ή Enhanced Relation Extraction: Significantly improved accuracy and robustness.
πŸ”Ή Improved Prompt Understanding: Optimized for open-information extraction tasks.
πŸ”Ή Better Named Entity Recognition (NER): More accurate and reliable results.

πŸ”§ How We Made It Better:
These advancements were made possible by:
πŸ”Ή Leveraging a better and more diverse dataset.
πŸ”Ή Using a larger backbone model for increased capacity.
πŸ”Ή Implementing advanced model merging techniques.
πŸ”Ή Employing self-learning strategies for continuous improvement.
πŸ”Ή Better training strategies and hyperparameters tuning.

πŸ“„ Read the Paper: https://arxiv.org/abs/2406.12925
βš™οΈ Try the Model: knowledgator/gliner-multitask-v1.0
πŸ’» Test the Demo: knowledgator/GLiNER_HandyLab
πŸ“Œ Explore the Repo: https://github.com/urchade/GLiNER
IhorΒ 
posted an update 5 months ago
view post
Post
397
πŸš€ Let’s transform LLMs into encoders πŸš€

Auto-regressive LMs have ruled, but encoder-based architectures like GLiNER are proving to be just as powerful for information extraction while offering better efficiency and interpretability. πŸ”βœ¨

Past encoder backbones were limited by small pre-training datasets and old techniques, but with innovations like LLM2Vec, we've transformed decoders into high-performing encoders! πŸ”„πŸ’‘

What’s New?
πŸ”ΉConverted Llama & Qwen decoders to advanced encoders
πŸ”ΉImproved GLiNER architecture to be able to work with rotary positional encoding
πŸ”ΉNew GLiNER (zero-shot NER) & GLiClass (zero-shot classification) models

πŸ”₯ Check it out:

New models: knowledgator/llm2encoder-66d1c76e3c8270397efc5b5e

GLiNER package: https://github.com/urchade/GLiNER

GLiClass package: https://github.com/Knowledgator/GLiClass

πŸ’» Read our blog for more insights, and stay tuned for what’s next!
https://medium.com/@knowledgrator/llm2encoders-e7d90b9f5966
IhorΒ 
posted an update 5 months ago
view post
Post
738
πŸš€ Meet the new GLiNER architecture πŸš€
GLiNER revolutionized zero-shot NER by demonstrating that lightweight encoders can achieve excellent results. We're excited to continue R&D with this spirit πŸ”₯. Our new bi-encoder and poly-encoder architectures were developed to address the main limitations of the original GLiNER architecture and bring the following new possibilities:

πŸ”Ή An unlimited number of entities can be recognized at once.
πŸ”ΉFaster inference when entity embeddings are preprocessed.
πŸ”ΉBetter generalization to unseen entities.

While the bi-encoder architecture can lack inter-label understanding, we developed a poly-encoder architecture with post-fusion. It achieves the same or even better results on many benchmarking datasets compared to the original GLiNER, while still offering the listed advantages of bi-encoders.
Now, it’s possible to run GLiNER with hundreds of entities much faster and more reliably.

πŸ“Œ Try the new models here:
knowledgator/gliner-bi-encoders-66c492ce224a51c54232657b
  • 4 replies
Β·
IhorΒ 
posted an update 7 months ago
view post
Post
893
πŸš€ Meet Our New Line of Efficient and Accurate Zero-Shot Classifiers! πŸš€

The new architecture brings better inter-label understanding and can solve complex classification tasks at a single forward pass.

Key Applications:
βœ… Multi-class classification (up to 100 classes in a single run)
βœ… Topic classification
βœ… Sentiment analysis
βœ… Event classification
βœ… Prompt-based constrained classification
βœ… Natural Language Inference
βœ… Multi- and single-label classification

knowledgator/gliclass-6661838823756265f2ac3848
knowledgator/GLiClass_SandBox
knowledgator/gliclass-base-v1.0-lw
IhorΒ 
posted an update 7 months ago
view post
Post
597
We’re thrilled to share our latest technical paper on the multi-task GLiNER model. Our research dives into the following exciting and forward-thinking topics:

πŸ” Zero-shot NER & Information Extraction: We demonstrate that with diverse and ample data, paired with the right architecture, encoders can achieve impressive results across various extraction tasks;

πŸ› οΈ Synthetic Data Generation: Leveraging open labelling by LLMs like Llama, we generated high-quality training data. Our student model even outperformed the teacher model, highlighting the potential of this approach.

πŸ€– Self-Learning: Our model showed consistent improvements in performance without labelled data, achieving up to a 12% increase in F1 score for initially challenging topics. This ability to learn and improve autonomously is a very perspective direction of future research!

GLiNER multi-task: Generalist Lightweight Model for Various Information Extraction Tasks (2406.12925)
knowledgator/gliner-multitask-large-v0.5
knowledgator/GLiNER_HandyLab


#!pip install gliner -U

from gliner import GLiNER

model = GLiNER.from_pretrained("knowledgator/gliner-multitask-large-v0.5")

text = """
Microsoft was founded by Bill Gates and Paul Allen on April 4, 1975 to develop and sell BASIC interpreters for the Altair 8800. 
"""

labels = ["founder", "computer", "software", "position", "date"]

entities = model.predict_entities(text, labels)

for entity in entities:
    print(entity["text"], "=>", entity["label"])