Commit 
							
							·
						
						97ba9b9
	
1
								Parent(s):
							
							6da9972
								
Fix typos in code sample (#1)
Browse files- Fix typos in code sample (074edeb01f6de68997bebccc25ff14c341edd6ed)
Co-authored-by: Sriram Krishna <[email protected]>
    	
        README.md
    CHANGED
    
    | @@ -35,17 +35,17 @@ Note that this model is primarily aimed at being fine-tuned on tasks that use th | |
| 35 | 
             
            You can use this model directly for masked language modeling:
         | 
| 36 |  | 
| 37 | 
             
            ```python
         | 
| 38 | 
            -
            from transformers import AutoTokenizer,  | 
| 39 | 
             
            tokenizer = AutoTokenizer.from_pretrained("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 40 | 
            -
            mlm_model =  | 
| 41 | 
             
            ```
         | 
| 42 |  | 
| 43 | 
             
            You can also fine-tune it for SequenceClassification, SequentialSentenceClassification, and MultipleChoice down-stream tasks:
         | 
| 44 |  | 
| 45 | 
             
            ```python
         | 
| 46 | 
            -
            from transformers import AutoTokenizer,  | 
| 47 | 
             
            tokenizer = AutoTokenizer.from_pretrained("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 48 | 
            -
            doc_classifier =  | 
| 49 | 
             
            ```
         | 
| 50 |  | 
| 51 | 
             
            ## Limitations and bias
         | 
|  | |
| 35 | 
             
            You can use this model directly for masked language modeling:
         | 
| 36 |  | 
| 37 | 
             
            ```python
         | 
| 38 | 
            +
            from transformers import AutoTokenizer, AutoModelForForMaskedLM
         | 
| 39 | 
             
            tokenizer = AutoTokenizer.from_pretrained("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 40 | 
            +
            mlm_model = AutoModelForMaskedLM("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 41 | 
             
            ```
         | 
| 42 |  | 
| 43 | 
             
            You can also fine-tune it for SequenceClassification, SequentialSentenceClassification, and MultipleChoice down-stream tasks:
         | 
| 44 |  | 
| 45 | 
             
            ```python
         | 
| 46 | 
            +
            from transformers import AutoTokenizer, AutoModelForSequenceClassification
         | 
| 47 | 
             
            tokenizer = AutoTokenizer.from_pretrained("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 48 | 
            +
            doc_classifier = AutoModelForSequenceClassification.from_pretrained("kiddothe2b/hierarchical-transformer-base-4096", trust_remote_code=True)
         | 
| 49 | 
             
            ```
         | 
| 50 |  | 
| 51 | 
             
            ## Limitations and bias
         | 
