Spaces:
Sleeping
Sleeping
Update model.py
Browse files
model.py
CHANGED
@@ -1,6 +1,8 @@
|
|
1 |
from transformers import pipeline
|
2 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
3 |
|
|
|
4 |
def modelFeedback(ats_score, resume_data, job_description):
|
5 |
"""
|
6 |
Generate ATS feedback by utilizing a pre-configured pipeline.
|
@@ -32,6 +34,15 @@ def modelFeedback(ats_score, resume_data, job_description):
|
|
32 |
|
33 |
|
34 |
# Load the tokenizer and model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-1B")
|
36 |
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B")
|
37 |
try:
|
|
|
1 |
from transformers import pipeline
|
2 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
3 |
+
import os
|
4 |
|
5 |
+
)
|
6 |
def modelFeedback(ats_score, resume_data, job_description):
|
7 |
"""
|
8 |
Generate ATS feedback by utilizing a pre-configured pipeline.
|
|
|
34 |
|
35 |
|
36 |
# Load the tokenizer and model
|
37 |
+
huggingface_token = os.environ.get("KEY2")
|
38 |
+
|
39 |
+
tokenizer = AutoTokenizer.from_pretrained(
|
40 |
+
"meta-llama/Llama-3.2-1B",
|
41 |
+
use_auth_token=huggingface_token
|
42 |
+
)
|
43 |
+
model = AutoModelForCausalLM.from_pretrained(
|
44 |
+
"meta-llama/Llama-3.2-1B",
|
45 |
+
use_auth_token=huggingface_token
|
46 |
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-1B")
|
47 |
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B")
|
48 |
try:
|