reach-vb HF staff commited on
Commit
8eb9656
·
1 Parent(s): a23d0da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -19,6 +19,9 @@ We further release a set of stereophonic capable models. Those were fine tuned f
19
  from the mono models. The training data is otherwise identical and capabilities and limitations are shared with the base modes. The stereo models work by getting 2 streams of tokens from the EnCodec model, and interleaving those using
20
  the delay pattern.
21
 
 
 
 
22
  MusicGen is a text-to-music model capable of genreating high-quality music samples conditioned on text descriptions or audio prompts.
23
  It is a single stage auto-regressive Transformer model trained over a 32kHz EnCodec tokenizer with 4 codebooks sampled at 50 Hz.
24
  Unlike existing methods, like MusicLM, MusicGen doesn't require a self-supervised semantic representation, and it generates all 4 codebooks in one pass.
@@ -79,11 +82,11 @@ import scipy
79
  import torch
80
  from transformers import pipeline
81
 
82
- synthesiser = pipeline("text-to-audio", "facebook/musicgen-stereo-small", device="cuda", torch_dtype=torch.float16)
83
 
84
- music = synthesiser("lo-fi music with a soothing melody", forward_params={"do_sample": True})
85
 
86
- scipy.io.wavfile.write("musicgen_out.wav", rate=music["sampling_rate"], music=audio["audio"])
87
  ```
88
 
89
  3. Run inference via the Transformers modelling code. You can use the processor + generate code to convert text into a mono 32 kHz audio waveform for more fine-grained control.
 
19
  from the mono models. The training data is otherwise identical and capabilities and limitations are shared with the base modes. The stereo models work by getting 2 streams of tokens from the EnCodec model, and interleaving those using
20
  the delay pattern.
21
 
22
+ Stereophonic sound, also known as stereo, is a technique used to reproduce sound with depth and direction.
23
+ It uses two separate audio channels played through speakers or headphones arranged so that it sounds like you're listening from different angles.
24
+
25
  MusicGen is a text-to-music model capable of genreating high-quality music samples conditioned on text descriptions or audio prompts.
26
  It is a single stage auto-regressive Transformer model trained over a 32kHz EnCodec tokenizer with 4 codebooks sampled at 50 Hz.
27
  Unlike existing methods, like MusicLM, MusicGen doesn't require a self-supervised semantic representation, and it generates all 4 codebooks in one pass.
 
82
  import torch
83
  from transformers import pipeline
84
 
85
+ synthesiser = pipeline("text-to-audio", "facebook/musicgen-stereo-small", device="cuda:0", torch_dtype=torch.float16)
86
 
87
+ music = synthesiser("lo-fi music with a soothing melody", forward_params={"max_new_tokens": 256})
88
 
89
+ sf.write("musicgen_out.wav", music["audio"][0].T, music["sampling_rate"])
90
  ```
91
 
92
  3. Run inference via the Transformers modelling code. You can use the processor + generate code to convert text into a mono 32 kHz audio waveform for more fine-grained control.