Tahsin Hasem commited on
Commit
896e66e
·
1 Parent(s): f43bb1f

Generate text before starting past api

Browse files
Files changed (1) hide show
  1. main.py +21 -2
main.py CHANGED
@@ -22,6 +22,7 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
22
 
23
  import transformers
24
  import torch
 
25
 
26
  model_id = "deepcogito/cogito-v1-preview-llama-3B"
27
 
@@ -33,6 +34,23 @@ pipeline = transformers.pipeline(
33
  )
34
 
35
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  app = FastAPI()
37
 
38
  class EchoMessage(BaseModel):
@@ -54,11 +72,12 @@ async def generate_text(item: Item):
54
  max_new_tokens=512,
55
  )
56
 
57
- print("request got")
58
 
59
  resp = outputs[0]["generated_text"][-1]
60
 
61
- print("Response generated", resp)
 
62
  return {"response": resp}
63
 
64
  @app.get("/")
 
22
 
23
  import transformers
24
  import torch
25
+ import logging
26
 
27
  model_id = "deepcogito/cogito-v1-preview-llama-3B"
28
 
 
34
  )
35
 
36
 
37
+ print("Pipeline loaded")
38
+ logging.info("Pipeline loaded")
39
+
40
+ messages = [
41
+ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
42
+ {"role": "user", "content": "Give me a short introduction to LLMs."},
43
+ ]
44
+
45
+ outputs = pipeline(
46
+ messages,
47
+ max_new_tokens=512,
48
+ )
49
+
50
+ logging.info("Generated text")
51
+ print(outputs[0]["generated_text"][-1])
52
+
53
+
54
  app = FastAPI()
55
 
56
  class EchoMessage(BaseModel):
 
72
  max_new_tokens=512,
73
  )
74
 
75
+ logging.info("request got")
76
 
77
  resp = outputs[0]["generated_text"][-1]
78
 
79
+ logging.info("Response generated")
80
+
81
  return {"response": resp}
82
 
83
  @app.get("/")