Spaces:
				
			
			
	
			
			
		Running
		
			on 
			
			Zero
	
	
	
			
			
	
	
	
	
		
		
		Running
		
			on 
			
			Zero
	Update
Browse files
    	
        app.py
    CHANGED
    
    | @@ -34,7 +34,7 @@ pose_image_processor = AutoProcessor.from_pretrained(pose_model_name) | |
| 34 | 
             
            pose_model = VitPoseForPoseEstimation.from_pretrained(pose_model_name, device_map=device)
         | 
| 35 |  | 
| 36 |  | 
| 37 | 
            -
            @spaces.GPU
         | 
| 38 | 
             
            @torch.inference_mode()
         | 
| 39 | 
             
            def process_image(image: PIL.Image.Image) -> tuple[PIL.Image.Image, list[dict]]:
         | 
| 40 | 
             
                inputs = person_image_processor(images=image, return_tensors="pt").to(device)
         | 
| @@ -107,6 +107,7 @@ def process_image(image: PIL.Image.Image) -> tuple[PIL.Image.Image, list[dict]]: | |
| 107 | 
             
                return vertex_annotator.annotate(scene=annotated_frame, key_points=keypoints), human_readable_results
         | 
| 108 |  | 
| 109 |  | 
|  | |
| 110 | 
             
            def process_video(
         | 
| 111 | 
             
                video_path: str,
         | 
| 112 | 
             
                progress: gr.Progress = gr.Progress(track_tqdm=True),  # noqa: ARG001, B008
         | 
| @@ -133,9 +134,6 @@ def process_video( | |
| 133 | 
             
                return out_file.name
         | 
| 134 |  | 
| 135 |  | 
| 136 | 
            -
            process_video.zerogpu = True  # type: ignore
         | 
| 137 | 
            -
             | 
| 138 | 
            -
             | 
| 139 | 
             
            with gr.Blocks(css_paths="style.css") as demo:
         | 
| 140 | 
             
                gr.Markdown(DESCRIPTION)
         | 
| 141 |  | 
|  | |
| 34 | 
             
            pose_model = VitPoseForPoseEstimation.from_pretrained(pose_model_name, device_map=device)
         | 
| 35 |  | 
| 36 |  | 
| 37 | 
            +
            @spaces.GPU(duration=5)
         | 
| 38 | 
             
            @torch.inference_mode()
         | 
| 39 | 
             
            def process_image(image: PIL.Image.Image) -> tuple[PIL.Image.Image, list[dict]]:
         | 
| 40 | 
             
                inputs = person_image_processor(images=image, return_tensors="pt").to(device)
         | 
|  | |
| 107 | 
             
                return vertex_annotator.annotate(scene=annotated_frame, key_points=keypoints), human_readable_results
         | 
| 108 |  | 
| 109 |  | 
| 110 | 
            +
            @spaces.GPU(duration=60)
         | 
| 111 | 
             
            def process_video(
         | 
| 112 | 
             
                video_path: str,
         | 
| 113 | 
             
                progress: gr.Progress = gr.Progress(track_tqdm=True),  # noqa: ARG001, B008
         | 
|  | |
| 134 | 
             
                return out_file.name
         | 
| 135 |  | 
| 136 |  | 
|  | |
|  | |
|  | |
| 137 | 
             
            with gr.Blocks(css_paths="style.css") as demo:
         | 
| 138 | 
             
                gr.Markdown(DESCRIPTION)
         | 
| 139 |  | 
