File size: 7,061 Bytes
133ccd4
 
 
96007f4
8453f63
 
 
133ccd4
f24d883
133ccd4
 
 
de46ee3
133ccd4
 
df03c6b
cef7dc5
e2f25e4
 
725cd35
5d8506a
e2f25e4
 
 
2ae94c0
93e4264
2ae94c0
 
5d8506a
8453f63
ed77379
8453f63
 
 
 
 
 
5b360d1
6d56961
 
b9dc011
 
4186f2a
b9dc011
ed77379
b9dc011
ed77379
b9dc011
 
 
d34cdb3
b9dc011
ed77379
b9dc011
 
6d56961
 
 
 
8ef0b3c
5d8506a
 
8ef0b3c
4186f2a
 
 
8ef0b3c
8453f63
7622f7a
8453f63
 
7622f7a
2ffbebd
c80330f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57db528
c80330f
2ffbebd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3dd0e68
2ffbebd
f6c94d2
 
96122eb
 
dd0af62
96122eb
 
 
0800621
e9fc459
c5ea46a
2ffbebd
6fd3683
2ffbebd
5b68d53
 
516d23a
 
 
 
 
 
 
 
 
 
 
 
 
 
c80330f
8453f63
 
df03c6b
 
 
a8c784b
df03c6b
 
 
3dd0e68
19faf7a
3dd0e68
 
de46ee3
 
23c274d
 
 
 
de46ee3
23c274d
 
de46ee3
e2f25e4
186f625
 
2ffbebd
efa809c
186f625
 
efa809c
186f625
de46ee3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
import argparse
import glob
import os.path

import torch
import torch.nn.functional as F

import gradio as gr

import onnxruntime as rt
import tqdm

from midi_synthesizer import synthesis
import TMIDIX

in_space = os.getenv("SYSTEM") == "spaces"
      
#=================================================================================================

@torch.no_grad()
def GenerateMIDI(progress=gr.Progress()):

    start_tokens = [3087, 3073+1, 3075+1]
    seq_len = 512
    max_seq_len = 2048
    temperature = 1.0
    verbose=False
    return_prime=False
    

    out = torch.FloatTensor([start_tokens])

    st = len(start_tokens)

    if verbose:
      print("Generating sequence of max length:", seq_len)

    progress(0, desc="Starting...")
    step = 0
    
    for i in progress.tqdm(range(seq_len)):

        try:
            x = out[:, -max_seq_len:]
    
            torch_in = x.tolist()[0]
    
            logits = torch.FloatTensor(session.run(None, {'input': [torch_in]})[0])[:, -1]

            probs = F.softmax(logits / temperature, dim=-1)
    
            sample = torch.multinomial(probs, 1)
    
            out = torch.cat((out, sample), dim=-1)

            if step % 16 == 0:
                print(step, '/', seq_len)

            step += 1
        
            if step >= seq_len:
                break
            
        except Exception as e:
            print('Error', e)
            break

    if return_prime:
      melody_chords_f = out[:, :]
    
    else:
      melody_chords_f = out[:, st:]

    melody_chords_f = melody_chords_f.tolist()[0]
    
    print('=' * 70)
    print('Sample INTs', melody_chords_f[:12])
    print('=' * 70)
    
    if len(melody_chords_f) != 0:
    
        song = melody_chords_f
        song_f = []
        time = 0
        dur = 0
        vel = 0
        pitch = 0
        channel = 0
    
        for ss in song:
    
          if ss > 0 and ss < 256:
    
              time += ss * 8
    
          if ss >= 256 and ss < 1280:
    
              dur = ((ss-256) // 8) * 32
              vel = (((ss-256) % 8)+1) * 15
    
          if ss >= 1280 and ss < 2816:
              channel = (ss-1280) // 128
              pitch = (ss-1280) % 128
    
              song_f.append(['note', int(time), int(dur), int(channel), int(pitch), int(vel) ])
    
    output_signature = 'Allegro Music Transformer'
    output_file_name = 'Allegro-Music-Transformer-Music-Composition'
    track_name='Project Los Angeles'
    list_of_MIDI_patches=[0, 24, 32, 40, 42, 46, 56, 71, 73, 0, 53, 19, 0, 0, 0, 0]
    number_of_ticks_per_quarter=500
    text_encoding='ISO-8859-1'
    
    output_header = [number_of_ticks_per_quarter, 
            [['track_name', 0, bytes(output_signature, text_encoding)]]]                                                    

    patch_list = [['patch_change', 0, 0, list_of_MIDI_patches[0]], 
                    ['patch_change', 0, 1, list_of_MIDI_patches[1]],
                    ['patch_change', 0, 2, list_of_MIDI_patches[2]],
                    ['patch_change', 0, 3, list_of_MIDI_patches[3]],
                    ['patch_change', 0, 4, list_of_MIDI_patches[4]],
                    ['patch_change', 0, 5, list_of_MIDI_patches[5]],
                    ['patch_change', 0, 6, list_of_MIDI_patches[6]],
                    ['patch_change', 0, 7, list_of_MIDI_patches[7]],
                    ['patch_change', 0, 8, list_of_MIDI_patches[8]],
                    ['patch_change', 0, 9, list_of_MIDI_patches[9]],
                    ['patch_change', 0, 10, list_of_MIDI_patches[10]],
                    ['patch_change', 0, 11, list_of_MIDI_patches[11]],
                    ['patch_change', 0, 12, list_of_MIDI_patches[12]],
                    ['patch_change', 0, 13, list_of_MIDI_patches[13]],
                    ['patch_change', 0, 14, list_of_MIDI_patches[14]],
                    ['patch_change', 0, 15, list_of_MIDI_patches[15]],
                    ['track_name', 0, bytes(track_name, text_encoding)]]

    output = output_header + [patch_list + song_f]

    midi_data = TMIDIX.score2midi(output, text_encoding)
    
    with open(f"Allegro-Music-Transformer-Music-Composition.mid", 'wb') as f:
        f.write(midi_data)

    output1 = []
    itrack = 1
    
    opus =  TMIDIX.score2opus(output)
          
    while itrack < len(opus):
        for event in opus[itrack]:
            if (event[0] == 'note_on') or (event[0] == 'note_off'): 
                output1.append(event)
        itrack += 1
        
    audio = synthesis([500, output1], 'SGM-v2.01-YamahaGrand-Guit-Bass-v2.7.sf2')

    midi_plot = []
    
    x = []
    y =[]
    c = []
    
    colors = ['red', 'yellow', 'green', 'cyan', 'blue', 'pink', 'orange', 'purple', 'gray', 'white', 'gold', 'silver']
    
    for s in song_f:
      x.append(s[1] / 1000)
      y.append(s[4])
      c.append(colors[s[3]])

    midi_plot.append([x, y, c])

    yield [500, output1], midi_plot, "Allegro-Music-Transformer-Music-Composition.mid", (44100, audio)
        
#=================================================================================================

if __name__ == "__main__":

    parser = argparse.ArgumentParser()
    parser.add_argument("--share", action="store_true", default=False, help="share gradio app")
    parser.add_argument("--port", type=int, default=7860, help="gradio server port")
    opt = parser.parse_args()

    print('Loading model...')
    session = rt.InferenceSession('Allegro_Music_Transformer_Small_Trained_Model_56000_steps_0.9399_loss_0.7374_acc.onnx', providers=['CUDAExecutionProvider', 'CPUExecutionProvider'])
    print('Done!')
    
    app = gr.Blocks()
    with app:
        gr.Markdown("<h1 style='text-align: center; margin-bottom: 1rem'>Allegro Music Transformer</h1>")
        gr.Markdown("![Visitors](https://api.visitorbadge.io/api/visitors?path=asigalov61.Allegro-Music-Transformer&style=flat)\n\n"
                    "Full-attention multi-instrumental music transformer featuring asymmetrical encoding with octo-velocity, and chords counters tokens, optimized for speed and performance\n\n"
                    "Check out [Allegro Music Transformer](https://github.com/asigalov61/Allegro-Music-Transformer) on GitHub!\n\n"
                    "[Open In Colab]"
                    "(https://colab.research.google.com/github/asigalov61/Allegro-Music-Transformer/blob/main/Allegro_Music_Transformer_Composer.ipynb)"
                    " for faster execution and endless generation"
                        )
                
        run_btn = gr.Button("generate", variant="primary")

        output_midi_seq = gr.Variable()
        output_plot = gr.ScatterPlot(label="output plot")
        output_audio = gr.Audio(label="output audio", format="mp3", elem_id="midi_audio")
        output_midi = gr.File(label="output midi", file_types=[".mid"])
        run_event = run_btn.click(GenerateMIDI, [], [output_midi_seq, output_plot, output_midi, output_audio])
        
        app.queue(2).launch(server_port=opt.port, share=opt.share, inbrowser=True)