playing tone with pyaudio - python

im trying to create an audio stream player that will receive audio through TCP, play it and save into file. This is just the player block part. For some reason pyaudio doesnt play the sine wave (sounddevice does, so its fine in the source format at least). Ive tried to play the converted format with sounddevice but without success, so that's probably where the issue lies. Can anyone point me to the right direction? thanks.
import numpy as np
import pyaudio
import sounddevice as sd
from array import array
import wave
sample_rate = 44100
Fs = 50000
f = 5000
sample = 150000
def play_stream(ob):
"converting object to pyauido required format"
tone = ob.astype(np.int16)
bytestream = tone.tobytes()
"playing data with pyaudio"
p = pyaudio.PyAudio()
stream = p.open(format=pyaudio.paInt16,
channels=1,
rate=sample_rate,
output=True)
if len(bytestream)>0:
stream.write(bytestream)
else:
print("no data!")
stream.stop_stream()
stream.close()
print("* stream finished")
p.terminate()
"testing data in alternate method"
sd.play(ob, 44100)
print ("play finished!")
"generating sine wave"
x = np.arange(sample)
ob = np.sin(2 * np.pi * f * x /Fs)
"playing sine wave"
play_stream(ob)
Edit: If i swap the format and data to float 32 it works. No idea why.

np.sin' returns a value in the range of -1 to 1. The 'astype' call is just casting it to a 16-bit integer which will give possible outputs of -1, 0, and 1. To use the full 16-bit range you need to multiply by 32767 before quantization.
scaled = ob * 32767
tone = scaled.totype(np.int16)

Related

Volume control in Pyaudio

I have pilfered some of the following code from previous stack exchange posts about
creating simple sine wave tones with pyaudio. The code is written to loop through the notes of the chromatic scale over two octaves. Each note has a frequency selected from the array named "notes". My questions: Why is the volume changing with each note? How can I keep the volume the same for each note? Please know that I am a pyaudio beginner.
import pyaudio
import numpy as np
import time
from numpy import zeros
p = pyaudio.PyAudio()
SR = 44100 # sampling rate
duration = 2.0 # seconds
durationPause = 2.0
#Frequency of pitches
notes = [220.00, 233.08, 246.94, 261.63, 277.18, 293.66, 311.13, 329.63, 349.23, 369.99, 392.00, 415.30, 440.00, 466.16, 493.88, 523.25, 554.37, 587.33, 622.25, 659.25, 698.46, 739.99, 783.99, 830.61, 880.00]
stream = p.open(format=pyaudio.paFloat32,
channels=1,
rate=SR,
output=True)
count = 0
while count < 23:
count += 1
f = notes[count]
#create sine wave of frequency f
samples = (np.sin(2.0*np.pi*np.arange(int(SR*duration))*(f/SR))).astype(np.float32)
#create pause
pause = zeros(int(SR*durationPause)).astype(np.float32)
#concatenate pitch and pause
samples = np.concatenate((samples, pause))
# Play sound.
stream.write((samples).tobytes())
stream.stop_stream()
stream.close()
p.terminate()

How to convert a numpy array to a mp3 file

I am using the soundcard library to record my microphone input, it records in a NumPy array and I want to grab that audio and save it as an mp3 file.
Code:
import soundcard as sc
import numpy
import threading
speakers = sc.all_speakers() # Gets a list of the systems speakers
default_speaker = sc.default_speaker() # Gets the default speaker
mics = sc.all_microphones() # Gets a list of all the microphones
default_mic = sc.get_microphone('Headset Microphone (Arctis 7 Chat)') # Gets the default microphone
# Records the default microphone
def record_mic():
print('Recording...')
with default_mic.recorder(samplerate=48000) as mic, default_speaker.player(samplerate=48000) as sp:
for _ in range(1000000000000):
data = mic.record(numframes=None) # 'None' creates zero latency
sp.play(data)
# Save the mp3 file here
recordThread = threading.Thread(target=record_mic)
recordThread.start()
With Scipy (to wav file)
You can easily convert to wav and then separately convert wav to mp3. More details here.
from scipy.io.wavfile import write
samplerate = 44100; fs = 100
t = np.linspace(0., 1., samplerate)
amplitude = np.iinfo(np.int16).max
data = amplitude * np.sin(2. * np.pi * fs * t)
write("example.wav", samplerate, data.astype(np.int16))
With pydub (to mp3)
Try this function from this excellent thread -
import pydub
import numpy as np
def write(f, sr, x, normalized=False):
"""numpy array to MP3"""
channels = 2 if (x.ndim == 2 and x.shape[1] == 2) else 1
if normalized: # normalized array - each item should be a float in [-1, 1)
y = np.int16(x * 2 ** 15)
else:
y = np.int16(x)
song = pydub.AudioSegment(y.tobytes(), frame_rate=sr, sample_width=2, channels=channels)
song.export(f, format="mp3", bitrate="320k")
#[[-225 707]
# [-234 782]
# [-205 755]
# ...,
# [ 303 89]
# [ 337 69]
# [ 274 89]]
write('out2.mp3', sr, x)
Note: Output MP3 will of cause be 16-bit, because MP3s are always 16 bit. However, you can set sample_width=3 as suggested by #Arty for 24-bit input.
As of now the accepted answer produces extremely distorted sound atleast in my case so here is the improved version :
#librosa read
y,sr=librosa.load(dir+file,sr=None)
y=librosa.util.normalize(y)
#pydub read
sound=AudioSegment.from_file(dir+file)
channel_sounds = sound.split_to_mono()
samples = [s.get_array_of_samples() for s in channel_sounds]
fp_arr = np.array(samples).T.astype(np.float32)
fp_arr /= np.iinfo(samples[0].typecode).max
fp_arr=np.array([x[0] for x in fp_arr])
#i normalize the pydub waveform with librosa for comparison purposes
fp_arr=librosa.util.normalize(fp_arr)
so you read the audiofile from any library and you have a waveform then you can export it to any pydub supported codec with this code below, i also used librosa read waveform and it works perfect.
wav_io = io.BytesIO()
scipy.io.wavfile.write(wav_io, sample_rate, waveform)
wav_io.seek(0)
sound = AudioSegment.from_wav(wav_io)
with open("file_exported_by_pydub.mp3",'wb') as af:
sound.export(
af,
format='mp3',
codec='mp3',
bitrate='160000',
)

Why is this just playing a quarter of a sine wave in pyaudio?

I'm trying to create a sine wave with pyaudio, and it seems like this code should create an entire waveform, repeated 1000 times:
import pyaudio
import numpy as np
p = pyaudio.PyAudio()
fs = 44100 # sampling rate, Hz, must be integer
f = 440.0 # sine frequency, Hz, may be float
duration = 1/f # in seconds, may be float
def sine_sampler(length, frequency, sampling_rate=44100):
"""Generate sine samples of volume 1"""
sine_samples = (np.sin(2 * np.pi * np.arange(length) * frequency / sampling_rate))
return sine_samples.astype(np.float32)
stream = p.open(format=pyaudio.paFloat32, channels=1, rate=fs, output=True)
sound = sine_sampler(duration * fs, f)
print(sound)
for i in range(1000):
stream.write(sound)
stream.stop_stream()
stream.close()
p.terminate()
When I print the sound after generating it, I get the expected behavior: an array that cycles up and down, as one cycle of a sine wave:
[ 0. 0.06264833 0.12505053 0.18696144 0.24813785 0.30833942
0.3673296 0.42487666 0.48075455 0.53474367 0.586632 0.6362156
0.6832998 0.72769946 0.76924026 0.807759 0.84310424 0.87513727
0.9037321 0.92877656 0.9501721 0.9678348 0.98169506 0.9916987
0.9978062 0.9999937 0.99825245 0.9925895 0.983027 0.9696024
0.9523687 0.9313933 0.90675884 0.87856203 0.84691364 0.811938
0.77377254 0.73256713 0.6884838 0.64169556 0.5923863 0.5407498
0.4869888 0.43131465 0.37394598 0.3151082 0.25503248 0.19395483
0.13211517 0.06975647 0.00712373 -0.055537 -0.11797953 -0.17995858
-0.24123062 -0.30155495 -0.36069456 -0.41841713 -0.47449586 -0.5287105
-0.5808479 -0.63070345 -0.67808115 -0.7227949 -0.76466894 -0.8035389
-0.83925205 -0.87166804 -0.90065956 -0.92611265 -0.94792736 -0.96601796
-0.9803134 -0.9907575 -0.99730927 -0.9999429 -0.9986481 -0.99342996
-0.98430896 -0.9713209 -0.9545169 -0.9339628 -0.90973955 -0.8819423
-0.85068005 -0.81607586 -0.77826554 -0.7373977 -0.6936328 -0.6471429
-0.59811056 -0.54672843 -0.4931984 -0.43773076 -0.3805434 -0.32186103
-0.2619142 -0.20093836 -0.1391731 -0.07686108 -0.0142471 ]
However, the resulting sine wave is only the first quarter of a sine wave, repeated again and again, as seen in this screenshot from Audacity:
It's only when I change duration to 4/f, getting four cycles, that it actually produces a smooth, complete sine wave.
Experiments with other durations support this same pattern of only playing part of the sound. Why is only the first fourth of each cycle being played?

How to change audio playback speed using Pydub?

I am new learner of audio editing libs - Pydub. I want to change some audio files' playback speed using Pydub(say .wav/mp3 format files), but I don't know how to make it. The only module I saw that could possibly deal with this problem is speedup module in effect.py. However, there is no explanation about how I am supposed to call it.
Could anyone kindly explain how to do this task in Pydub? Many thanks!
(A related question: Pydub - How to change frame rate without changing playback speed, but what I want to do is to change the playback speed without changing the audio quality.)
This can be done using pyrubberband package which requires rubberband library that can stretch audio while keeping the pitch and high quality. I was able to install the library on MacOS using brew, and same on Ubuntu with apt install. For extreme stretching, look into PaulStretch
brew install rubberband
This works simply with librosa package
import librosa
import pyrubberband
import soundfile as sf
y, sr = librosa.load(filepath, sr=None)
y_stretched = pyrubberband.time_stretch(y, sr, 1.5)
sf.write(analyzed_filepath, y_stretched, sr, format='wav')
To make pyrubberband work directly with AudioSegment from pydub without librosa I fiddled this function:
def change_audioseg_tempo(audiosegment, tempo, new_tempo):
y = np.array(audiosegment.get_array_of_samples())
if audiosegment.channels == 2:
y = y.reshape((-1, 2))
sample_rate = audiosegment.frame_rate
tempo_ratio = new_tempo / tempo
print(tempo_ratio)
y_fast = pyrb.time_stretch(y, sample_rate, tempo_ratio)
channels = 2 if (y_fast.ndim == 2 and y_fast.shape[1] == 2) else 1
y = np.int16(y_fast * 2 ** 15)
new_seg = pydub.AudioSegment(y.tobytes(), frame_rate=sample_rate, sample_width=2, channels=channels)
return new_seg
from pydub import AudioSegment
from pydub import effects
root = r'audio.wav'
velocidad_X = 1.5 # No puede estar por debajo de 1.0
sound = AudioSegment.from_file(root)
so = sound.speedup(velocidad_X, 150, 25)
so.export(root[:-4] + '_Out.mp3', format = 'mp3')
I know it's late but I wrote a program to convert mp3 to different playback speed.
First, Convert the .MP3 -> .Wav because PYRubberBand supports only .wav format. Then streach the time and pitch at the same time to avoid chipmunk effect.
import wave
import sys
from pydub import AudioSegment
#sound = AudioSegment.from_file("deviprasadgharpehai.mp3")
sound = AudioSegment.from_mp3(sys.argv[1])
sound.export("file.wav", format="wav")
print(sys.argv[1])
import soundfile as sf
import pyrubberband as pyrb
y, sr = sf.read("file.wav")
# Play back at extra low speed
y_stretch = pyrb.time_stretch(y, sr, 0.5)
# Play back extra low tones
y_shift = pyrb.pitch_shift(y, sr, 0.5)
sf.write("analyzed_filepathX5.wav", y_stretch, sr, format='wav')
sound = AudioSegment.from_wav("analyzed_filepathX5.wav")
sound.export("analyzed_filepathX5.mp3", format="mp3")
# Play back at low speed
y_stretch = pyrb.time_stretch(y, sr, 0.75)
# Play back at low tones
y_shift = pyrb.pitch_shift(y, sr, 0.75)
sf.write("analyzed_filepathX75.wav", y_stretch, sr, format='wav')
sound = AudioSegment.from_wav("analyzed_filepathX75.wav")
sound.export("analyzed_filepathX75.mp3", format="mp3")
# Play back at 1.5X speed
y_stretch = pyrb.time_stretch(y, sr, 1.5)
# Play back two 1.5x tones
y_shift = pyrb.pitch_shift(y, sr, 1.5)
sf.write("analyzed_filepathX105.wav", y_stretch, sr, format='wav')
sound = AudioSegment.from_wav("analyzed_filepathX105.wav")
sound.export("analyzed_filepathX105.mp3", format="mp3")
# Play back at same speed
y_stretch = pyrb.time_stretch(y, sr, 1)
# Play back two smae-tones
y_shift = pyrb.pitch_shift(y, sr, 1)
sf.write("analyzed_filepathXnormal.wav", y_stretch, sr, format='wav')
sound = AudioSegment.from_wav("analyzed_filepathXnormal.wav")
sound.export("analyzed_filepathXnormal.mp3", format="mp3")
**Make Sure to install **
Wave, AudioSegment, FFmpeg, PYRubberBand, Soundfile
To use this Run,
python3 filename.py mp3filename.mp3
To change the speed of the audio without changing the pitch (or creating chipmunk effect). You can use below code.
from pydub import AudioSegment
from pydub.effects import speedup
audio = AudioSegment.from_mp3(song.mp3)
new_file = speedup(audio,1.5,150)
new_file.export("file.mp3", format="mp3")

How to change audio speed without changing pitch?

I need to apply audio to video at certain time with certain duration, but some audio duration is bigger(or smaller) then needed. How to change speed of audio without changing pitch? I tried to change fps(by multiplying to division of needed duration to audio duration) but it is not work as I want.
original = VideoFileClip("orig.mp4")
clips = [orig.audio.volumex(0.3)]
subs = [] #some array
i = 0
for sub in subs:
clip = AudioFileClip("\\temp{}.mp3")
mult = clip.duration / (sub.end - sub.start) + 0.00001
clip = AudioArrayClip(clip.to_soundarray(buffersize=500, fps=24000/mult), fps=24000).set_start(sub.start).set_end(sub.end)
clips.append(clip)
i += 1
final = CompositeAudioClip(clips)
final.write_audiofile("final.mp3")
you can use librosa module:
from scipy.io import wavfile
import librosa, numpy as np
song, fs = librosa.load("song.wav")
song_2_times_faster = librosa.effects.time_stretch(song, 2)
scipy.io.wavfile.write("song_2_times_faster.wav", fs, song_2_times_faster) # save the song
Using wave: Change the sampling rate
import wave
CHANNELS = 1
swidth = 2
Change_RATE = 2
spf = wave.open('VOZ.wav', 'rb')
RATE=spf.getframerate()
signal = spf.readframes(-1)
wf = wave.open('changed.wav', 'wb')
wf.setnchannels(CHANNELS)
wf.setsampwidth(swidth)
wf.setframerate(RATE*Change_RATE)
wf.writeframes(signal)
wf.close()

Categories