metadata
language: de
datasets:
- news_commentary
widget:
- text: Unberechenbar, gefährlich, ja, auf jeden Fall.
example_title: Fluent example 1
- text: Aber hinterher... oh, oh...
example_title: Fluent example 2
- text: Nettes Haus, was? - Ja.
example_title: Fluent example 3
- text: Wissqween Sisssasde, adddddqwe12was Mdddilednberg war, 122huh?
example_title: Disfluent example 1
- text: asdaojn;klL:JjJALSJD
example_title: Disfluent example 2
- text: Was dDadasdDasein erster aaaaEind2ruck?
example_title: Disfluent example 3
license: other
This model was trained for evaluating linguistic acceptability and grammaticality. The finetuning was carried out based off the bert-base-german-cased.
To use the model:
from transformers import pipeline
classifier = pipeline("text-classification", model = 'EIStakovskii/bert-base-german-cased_fluency')
print(classifier("Wissqween Sisssasde, adddddqwe12was Mdddilednberg war, 122huh?"))
Label_1 means ACCEPTABLE - the sentence is perfectly understandable by native speakers and has no serious grammatic and syntactic flaws.
Label_0 means NOT ACCEPTABLE - the sentence is flawed both orthographically and grammatically.
The model was trained on 50 thousand German sentences from the news_commentary dataset. Out of 50 thousand 25 thousand sentences were algorithmically corrupted using the open source Python library. The library was originally developed by aylliote, but it was slightly adapted for the purposes of this model.