robinwitch commited on
Commit
bd336d0
·
1 Parent(s): f218e94
Files changed (1) hide show
  1. app.py +14 -13
app.py CHANGED
@@ -106,19 +106,19 @@ class BaseTrainer(object):
106
 
107
  self.checkpoint_path = tmp_dir
108
  args.tmp_dir = tmp_dir
109
- if self.rank == 0:
110
- self.test_data = __import__(f"dataloaders.{args.dataset}", fromlist=["something"]).CustomDataset(args, "test")
111
- self.test_loader = torch.utils.data.DataLoader(
112
- self.test_data,
113
- batch_size=1,
114
- shuffle=False,
115
- num_workers=args.loader_workers,
116
- drop_last=False,
117
- )
118
  logger.info(f"Init test dataloader success")
119
- model_module = __import__(f"models.{args.model}", fromlist=["something"])
120
-
121
- self.model = getattr(model_module, args.g_name)(args).cuda()
122
 
123
  if self.rank == 0:
124
  logger.info(self.model)
@@ -768,7 +768,8 @@ demo = gr.Interface(
768
  ],
769
  title='SynTalker: Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation',
770
  description="1. Upload your audio. <br/>\
771
- 2. Then, sit back and wait for the rendering to happen! This may take a while (e.g. 2 minutes) <br/>\
 
772
  3. After, you can view the videos. <br/>\
773
  4. Notice that we use a fix face animation, our method only produce body motion. <br/>\
774
  5. Use DDPM sample strategy will generate a better result, while it will take more inference time. \
 
106
 
107
  self.checkpoint_path = tmp_dir
108
  args.tmp_dir = tmp_dir
109
+
110
+ self.test_data = __import__(f"dataloaders.{args.dataset}", fromlist=["something"]).CustomDataset(args, "test")
111
+ self.test_loader = torch.utils.data.DataLoader(
112
+ self.test_data,
113
+ batch_size=1,
114
+ shuffle=False,
115
+ num_workers=args.loader_workers,
116
+ drop_last=False,
117
+ )
118
  logger.info(f"Init test dataloader success")
119
+
120
+ from models.denoiser import MDM
121
+ self.model = MDM(args).cuda()
122
 
123
  if self.rank == 0:
124
  logger.info(self.model)
 
768
  ],
769
  title='SynTalker: Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation',
770
  description="1. Upload your audio. <br/>\
771
+ 2. Then, sit back and wait for the rendering to happen! This may take a while (e.g. 2-5 minutes) <br/>\
772
+ (The reason of running time so long is that provided GPU have an limitation in GPU running time, we must use CPU to handle some GPU tasks)\
773
  3. After, you can view the videos. <br/>\
774
  4. Notice that we use a fix face animation, our method only produce body motion. <br/>\
775
  5. Use DDPM sample strategy will generate a better result, while it will take more inference time. \