Lab 2: Build a synthesizer#

Last updated May 7, 2024

00. Content #

Mathematics#

  • quantifying rounding error

  • definition of the cent

  • rational approximations to irrational numbers

Programming Skills#

  • specialized flow structures (try/finally, for/else)

  • buffers

Embedded Systems#

  • the I2S protocol

  • blocking code

0. Required Hardware #

Write your name and email below:

Name: me

Email: me @purdue.edu

import matplotlib.pyplot as plt
import numpy as np

1. Intro #

One of the fundamental things about sounds is the timing. People are very sensitive to small differences in timing of audio signals. That is a big problem if you want to use a general-purpose computer to create or process sounds in real time, since a computer has many tasks to keep track of like monitoring network connections, checking for user input, managing file systems, moving programs into and out of working memory, health check to name a few. And those are just the things it does in the background, let alone if you ask it to do something as well! Most computers you can buy, and the smart phones today, contain specialized electronic chips for processing audio, freeing up the general-purpose parts to do other things.

Most of the complex algorithms that are performed on audio are done on DSP chips. They perform complex computations in a short amout of time. But most of them are very expensive This is where a microcontroller is an ideal choice: we get some of the flexibility of a general-purpose computer, while still being cheaper and keeping tight control over the timing of the program.

As an introduction to the tools of working with sound in real time, we will build a simple synthesizer – that is, an electronic piano.

2. The I2S protocol #

It should be no surprise that there is a special digital protocol used for audio signals. Other protocols, like I2C or SPI, simply add too much latency. The I2S protocol is much simpler than either I2C or SPI, and therefore can get much lower latency and higher data throughput. It consists of just three wires: one for a clock, one for data, and one for indicating whether the data should go to the left or right audio channel.

Today, we are going to use a module which accepts I2S input and drives a speaker directly. Wire it up as indicated in the following tables. If you are not sure you have done it correctly, then ask an instructor to verify your wiring is correct before plugging your circuit into the USB port.

MAX 98357

Speaker

-

(either wire)

+

(the other wire)

MAX 98357

Pico

What it does

Vin

VBUS

USB power directly to the speaker driver

GND

GND

A shared ground

BCLK

GP16

clock signal

LRC

GP17

left/right channel select

DIN

GP18

data signal

an image of the wiring described above

Your breadboard should look something like the picture above.

Test your circuit#

Once you are sure that your circuit is wired safely, open up Thonny and run the script circuittest.py on your Pico. You should hear a simple tone play for two seconds, and then it should stop. If your circuit does not do this, figure out what is going wrong before you continue. Your instructor may be able to help you narrow down the problem.

Let’s understand what is happening in this file, line by line.

"""
This code plays a simple tone to a connected I2S speaker
for two seconds second, just to see if everything is working.
"""

Every good Python program should have a docstring which summarizes what the program is for.

import time
from machine import I2S, Pin

This script only uses built-in modules, the time module to count out two seconds and the machine module to interface with hardware.

tone = b'\x00\x00\x05\x08\xea\x0f\x8f\x17\xd5\x1e\x9e%\xcf+P1\t6\xe89\xde<\xdd>\xdf?\xdf?\xdd>\xde<\xe89\t6P1\xcf+\x9e%\xd5\x1e\x8f\x17\xea\x0f\x05\x08\x00\x00\xfb\xf7\x16\xf0q\xe8+\xe1b\xda1\xd4\xb0\xce\xf7\xc9\x18\xc6"\xc3#\xc1!\xc0!\xc0#\xc1"\xc3\x18\xc6\xf7\xc9\xb0\xce1\xd4b\xda+\xe1q\xe8\x16\xf0\xfb\xf7'

This is the data which makes up the sound we play. We will analyze this in more detail shortly.

try:
    # [...]
finally:
    audio_out.deinit()

When an I2S object is created, it configures some parts of the microcontroller to run independently of the main process. It is important that these are stopped when we are done with them, even if there is an exception which makes our code stop. By using a try/finally block, we ensure that the code in the finally block runs no matter what.

audio_out = I2S(
    0,
    sck=Pin(16),
    ws=Pin(17),
    sd=Pin(18),
    mode=I2S.TX,
    bits=16,
    format=I2S.MONO,
    rate=22_050,
    ibuf=2000,
)

We make a new I2S object, telling it which pins to connect to, to start in transmitting mode, and a few other configuration parameters.

starttime = time.ticks_ms()
while time.ticks_ms() < starttime + 2000:
    audio_out.write(tone)

We figure out the current time, in milliseconds. We continually write the data tone to the I2S device, stopping only when at least 2000 milliseconds have passed since we started.

By using built in Micropython modules, we were able to get sound with not-too-many-lines of code.

Understanding the I2S object#

When we made the I2S object in the example code, we told it a few things. We told it to transmit (mode=I2S.TX), we told it to use 16 bits for each sample (bits=16), we told it to use mono rather than stereo sound (format=I2S.MONO), and we told it to use a sample rate of 22,050 Hz (rate=22_050). You likely remember that most audio equipment uses a sample rate of 44,100 Hz. We use a lower sample rate because our speaker is not sensitive enough to reproduce the higher frequencies 44,100 Hz would give us, and a lower sampling rate means less work for the microcontroller. Finally, we tell the I2S object to make a buffer of 2000 bytes (ibuf=2000) to hold audio in a queue, sent out at a controlled rate of 22,050 samples (44,100 bytes) per second. In effect, that means the buffer holds

\[ 2000\text{ bytes} \times \frac{1\text{ sample}}{2\text{ bytes}}\times \frac{1\text{ second}}{22,050\text{ samples}}\times\frac{1000\text{ milliseconds}}{1\text{ second}} = 45\text{ milliseconds} \]

of audio data.

Whenever we call audio_out.write(tone) the I2S object checks whether there is enough room in the buffer to hold tone, which is 100 bytes long. If there is enough room, it places tone into the buffer. Otherwise, it sits and waits for the room to be available, then writes tone to the buffer.

What kind of data is tone, exactly? We already said it consists of 2-byte samples. Let’s use numpy here to read it out. We will tell numpy to use the most common 2-byte data format, a 16 bit integer.

tone = b'\x00\x00\x05\x08\xea\x0f\x8f\x17\xd5\x1e\x9e%\xcf+P1\t6\xe89\xde<\xdd>\xdf?\xdf?\xdd>\xde<\xe89\t6P1\xcf+\x9e%\xd5\x1e\x8f\x17\xea\x0f\x05\x08\x00\x00\xfb\xf7\x16\xf0q\xe8+\xe1b\xda1\xd4\xb0\xce\xf7\xc9\x18\xc6"\xc3#\xc1!\xc0!\xc0#\xc1"\xc3\x18\xc6\xf7\xc9\xb0\xce1\xd4b\xda+\xe1q\xe8\x16\xf0\xfb\xf7'
samples = np.frombuffer(tone, dtype=np.int16)
print(samples)
[     0   2053   4074   6031   7893   9630  11215  12624  13833  14824
  15582  16093  16351  16351  16093  15582  14824  13833  12624  11215
   9630   7893   6031   4074   2053      0  -2053  -4074  -6031  -7893
  -9630 -11215 -12624 -13833 -14824 -15582 -16093 -16351 -16351 -16093
 -15582 -14824 -13833 -12624 -11215  -9630  -7893  -6031  -4074  -2053]

We can see that is a list of 50 integers, ranging from around -16,000 to 16,000. Let’s plot them to get an idea for what it is.

plt.plot(samples)
plt.show()
../../../_images/aa6d3672de3b817d5d762a619864e2f0a84d5cca3902933c8334e004285debb9.png

That looks like a clean sine wave! This one has been carefully constructed so that when it is repeated, there is a smooth transition from one sample to the next.

plt.plot(np.tile(samples, 3))
plt.show()
../../../_images/6d8a4dfced29723cefd67d31ffa0a803f79563f1f63ea3f1ab486514f112889c.png

Exercise 1 (20 pts)#

You know that each number in tone is one sample, and you know the sample rate. Use this information to plot tone object given below with units of milliseconds on the x-axis. Be sure to label the x-axis in your plot with the correct units.

tone = b'\x00\x00\x05\x08\xea\x0f\x8f\x17\xd5\x1e\x9e%\xcf+P1\t6\xe89\xde<\xdd>\xdf?\xdf?\xdd>\xde<\xe89\t6P1\xcf+\x9e%\xd5\x1e\x8f\x17\xea\x0f\x05\x08\x00\x00\xfb\xf7\x16\xf0q\xe8+\xe1b\xda1\xd4\xb0\xce\xf7\xc9\x18\xc6"\xc3#\xc1!\xc0!\xc0#\xc1"\xc3\x18\xc6\xf7\xc9\xb0\xce1\xd4b\xda+\xe1q\xe8\x16\xf0\xfb\xf7'

What is the frequency of this sine wave? Explain in 2-3 sentences how to compute it.

Write Answers for Exercise 1 Below

3. Designing good samples #

Remember that our goal is to make a synthesizer, so we should figure out how to make sounds of any frequency. We already know the data format we need – just a bunch of 16 bit integers. However, there are some subtleties which we need to address. In order to make them clear, let’s see what happens if we use a naive approach.

Let’s make some data which gives precisely a middle F, so \(440\times 2^{-1/3}\) Hz. We will start by doing an arbitrary number of samples, say, 200.

t = np.arange(200) / 22_050
middle_f_frequency = 440 * (2 ** (-1 / 3))
print(f'middle f has a frequency of {middle_f_frequency} Hz')
middle_f = np.sin(2 * np.pi * middle_f_frequency * t)

plt.plot(t, middle_f)
plt.show()
middle f has a frequency of 349.2282314330039 Hz
../../../_images/5ec123cd9704780e3cf72ff0ee6dbd290549852cd5984f054464be8bfadd24f3.png

This is a nice sine wave, but there is a problem which comes up if we try to repeat it.

plt.plot(np.tile(middle_f, 3))
plt.show()
../../../_images/7feab71c5a0ad2374c81f037fdb7e0bda5f5e2b48f81374700f926cc10c7a347.png

As you can see, that makes a nasty jump at every repeat. When we play this as audio, it makes a nasty buzz.

from IPython.display import Audio

Audio(np.tile(middle_f, 500), rate=22050)

Let’s be smarter about how we choose our number of samples. Let’s try and include just one period of the wave. One period is

\[ \frac{1}{440\times 2^{-1/3} \text{ Hz}} = 2.86\text{ milliseconds} \]

and there are 22,050 samples per second, so the ideal period is

\[ \frac{1}{440\times 2^{-1/3} \text{ Hz}} \times \frac{22,050\text{ samples}}{1\text{ second}} = 63.14\text{ samples} \]

Here is our first serious problem: that’s not a whole number of samples. Let’s hear what it sounds like if we round to the nearest number of samples.

plt.plot(np.tile(middle_f[:63], 3))
plt.show()

Audio(np.tile(middle_f[:63], 2000), rate=22050)
../../../_images/ba3c90e6843adc9a7307e89da30ca752eb6dc9f169216ba9723c9fad2acb9ae7.png

That is surely a lot better, but if you look closely at the graph you can still see a jump, and if you listen closely you can still hear a buzz.

The solution is to be less picky about our frequency. We wanted something with exactly \(440\times 2^{-1/3}\) Hz, but nobody’s ears could tell if it were off by a little bit. Instead of choosing the number of samples from the frequency, what if we choose a frequency from the number of samples? If we do a sound with exactly 63 samples per second, what frequency would that be? Simply

\[ \frac{1}{63\text{ samples}}\times \frac{22,050 \text{ samples}}{\text{second}} = 350\text{ Hz}, \]

which is pretty close to \(440\times 2^{-1/3} = 349.2\) Hz. Hear the difference for yourself. The buzzing is gone, but the pitch is slightly different. It is very close, so you may not be able to quite make it out.

t = np.arange(30_000)/22_050
true_f = np.sin(2*np.pi*middle_f_frequency*t)
Audio(true_f, rate=22050)
approximate_f = np.tile(np.sin(np.linspace(0, 2*np.pi, 63, endpoint=False)), 500)
Audio(approximate_f, rate=22050)

Cents for errors in frequency#

We found the 63-sample approximation is off by 0.8 Hz from the true frequency. How big of a deal is that? For discussing small differences in frequency, we usually use the unit called the cent. Each note on the keyboard is separated by 100 cents, hence the name. Since the notes are defined by ratios, so are cents.

The difference in frequency between two frequencies \(a\) and \(b\) is defined to be

\[ n=1200\log_2 \frac{b}{a}. \]

So, in our example, the approximation 350 Hz is

\[ 1200\log_2 \frac{350}{440\times 2^{-1/3}} = 3.82\text{ cents} \]

higher than a true middle F. People generally have a hard time distinguishing between notes which are within about 10 cents, though some people can notice differences of about 5 cents. That tells us that our approximation is certainly good enough.

Summary thus far#

We noticed that, in order to have a repeating audio clip which gives the exactly correct frequency for a given note, we would have jumps in between repeats which make a nasty buzzing sound. In order to get rid of the buzzing, we can change the frequency a little bit to make an audio clip with a nearby frequency which does not buzz but is off by a little bit. In the case of middle F, the true note has a frequency of 349.2 Hz, which we can approximate with a clip of 63 samples which has a frequency of 350 Hz.

Troublesome notes#

In a way we got lucky with middle F, because we found a true F has 63.14 samples/cycle which is very close to 63 samples/cycle. With some notes we are not so lucky. Consider, for example, middle B. This has a frequency of \(440\times 2^{1/6}=493.9\) Hz. That means having

\[ \frac{22,050 \text{ samples}}{\text{second}}\times\frac{1}{440\times 2^{1/6}} = 44.65\text{ samples} \]

in each cycle. We could have to round that pretty far to get to 45 samples. As a result, a rounded clip which repeats every 45 samples would have a frequency of

\[ \frac{22,050\text{ samples}}{\text{second}}\times\frac{1}{45\text{ samples}} = 490\text{ Hz}, \]

which is

\[ 1200\log_2\frac{490}{493.9} = -13.7\text{ cents} \]

off from the true frequency. That is easily audible, and a bit of a problem. How can we fix it?

44.65 is far from either 44 or 45, but it is pretty close to \(44+\frac{2}{3}\). If we had a clip that was \(3\times(44+\frac{2}{3})=134\) samples long, we could put three cycles of the note into our audio clip. That would give us a frequency of

\[ \frac{3\text{ cycles}}{134\text{ samples}}\times \frac{22,050\text{ samples}}{\text{second}}= 493.6 \text{ Hz}. \]

That is much closer to the true value, only off by 0.79 cents.

The rest of them#

We can use pandas to easily make a table of the frequencies from middle B to middle C. We can see just how good (or how bad) the simplest approximation (rounding to the nearest number of samples per audio clip) is.

import pandas as pd

notes = "C C# D D# E F F# G G# A A# B"
scale = pd.Series(440*2**((np.arange(12)-9)/12), index = notes.split())
samples_per_cycle = 22050/scale
rounded = np.rint(samples_per_cycle).astype(int)
approximation = 22050/rounded
errors = 1200*np.log2(approximation/scale)

summary = pd.DataFrame({
    'Correct frequency': scale,
    'Samples per cycle': samples_per_cycle,
    'Rounded': rounded,
    'Approximation frequency': approximation,
    'Difference (cents)': errors,
})
summary
Correct frequency Samples per cycle Rounded Approximation frequency Difference (cents)
C 261.625565 84.280754 84 262.500000 5.776679
C# 277.182631 79.550439 80 275.625000 -9.756128
D 293.664768 75.085616 75 294.000000 1.975158
D# 311.126984 70.871384 71 310.563380 -3.138957
E 329.627557 66.893679 67 329.104478 -2.749442
F 349.228231 63.139225 63 350.000000 3.821678
F# 369.994423 59.595493 60 367.500000 -11.711129
G 391.995436 56.250655 56 393.750000 7.731680
G# 415.304698 53.093548 53 416.037736 3.053041
A 440.000000 50.113636 50 441.000000 3.930158
A# 466.163762 47.300974 47 469.148936 11.050964
B 493.883301 44.646174 45 490.000000 -13.666129

Exercise 2 (30 pts)#

Six of the notes above are off by more than 5 cents from the correct frequency, when you just round to the nearest number of samples per cycle. We showed how you can make an audio clip with more cycles to get a better approximation. Use that same strategy to find a number of samples which approximates every note above, with an error of less than 5 cents.

Write Answers for Exercise 2 Below

4. The completed synthesizer #

Now that we know how many samples to use for each note, we can have the microcontroller play them when we press buttons. In order to make it easier to wire up a keypad worth of buttons, a module has been provided which simply connects 12 buttons to 12 GPIO pins on the Pico, connecting the pins to ground whenever the corresponding button is pressed. You could just as well have wired up 12 buttons individually yourself, but this module makes it all tidier.

Connect the module to GPIO pins 4-15, as in the picture below.

a picture showing how to connect the keypad module

Then, load up the script synth.py and load it in Thonny. When running it, you should find that you have a synthesizer, though it only plays the notes F and B.

Those are also the only two entries in the dictionary tones which are not silence. In Exercise 3 you will have to create audio clips to fill in the other entries. First, let’s see how the rest of the script works, and how the example tones were generated.

The script#

Let’s go through this script line by line.

"""
This is the driver code for a synthsizer.
"""
from machine import I2S, Pin

Every good Python file has a docstring. The only modules we are using are built in.

keys = {
    letter: Pin(i, Pin.IN, Pin.PULL_UP)
    for i, letter in zip(range(4, 16), "C C# D D# E F F# G G# A A# B".split())
}

This is a dictionary comprehension, a short way of writing a dictionary. The dictionary keys are letters (names of notes) and the values are pins. We configure the pins to use pull up resistors, so that when a button connects the corresponding pin to ground, the voltage changes.

silence = bytes(100)

This makes a byte string of all zeros:

bytes(100)
b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00'

This would be interpreted by the I2S object as 50 samples of zero, which is 2.3 milliseconds of silence.

tones = {
    #[...]
}

This is the dictionary which holds the audio clips. We will explore how these are made in a bit.

try:
    audio_out = I2S(
        # ...
    )

    while True:
        # ...
finally:
    audio_out.deinit()

The I2S object is configured in exactly the same way as in the previous example. The try/finally block is extra important here, because we are using a while True loop. The only way we will get out of that loop is if an exception is raised, and we want to be sure that audio_out.deinit() is called no matter what.

for letter, pin in keys.items():
    if not pin.value():
        audio_out.write(tones[letter])
        break
else:
    audio_out.write(silence)

Here we are using a somewhat uncommon structure, the for/else block. Notice that the else is not connected to the if, it’s connected to the for. The way this work is that the code in the else block only runs if the for loop does not end with a break. That is, a break statement would exit the for loop and skip the else block but if the for loop gets all the way through without a break then the else block runs.

The effect here is that the first pin we find pressed will get its corresponding tone written to the speaker buffer, and then we leave the loop. If we don’t find any button pressed, then we enter the else block and write 2.3 milliseconds of silence to the buffer instead.

Since this is in a loop, it will immediately run again and again, so we get a near-instant response when the buttons change.

Making the clips#

The byte strings in the tones dictionary were made using the following code, then copy/pasted into synth.py.

f = 2**14*np.sin(np.linspace(0, 2*np.pi, 63, endpoint=False))
plt.plot(f)
plt.show()
print(f.astype(np.int16).tobytes())
../../../_images/9468b268bd5c19d29819c7b70edd0f2c41ad0230c6322545d0537935422b89c1.png
b'\x00\x00_\x06\xae\x0c\xdd\x12\xdc\x18\x9b\x1e\r$#)\xd0-\t2\xc25\xf38\x93;\x9b=\x07?\xd2?\xfa?\x80?e>\xaa<V:l7\xf63\xfc/\x87+\xa4&_!\xc4\x1b\xe3\x15\xca\x0f\x89\t0\x03\xd0\xfcw\xf66\xf0\x1d\xea<\xe4\xa1\xde\\\xd9y\xd4\x04\xd0\n\xcc\x94\xc8\xaa\xc5V\xc3\x9b\xc1\x80\xc0\x06\xc0.\xc0\xf9\xc0e\xc2m\xc4\r\xc7>\xca\xf7\xcd0\xd2\xdd\xd6\xf3\xdbe\xe1$\xe7#\xedR\xf3\xa1\xf9'
b = 2**14*np.sin(np.linspace(0, 6*np.pi, 134, endpoint=False))
plt.plot(b)
plt.show()
print(b.astype(np.int16).tobytes())
../../../_images/c27ebeee59a85b17f76829a597909dd2b5cb3056db998e0f49db1ef8db42e1e5.png
b'\x00\x00\xf9\x08\xc4\x116\x1a$"d)\xd3/Q5\xc19\x0c=#?\xfb?\x8f?\xe2=\xfb:\xeb6\xc41\xa2+\xa3$\xeb\x1c\xa1\x14\xee\x0b\xff\x02\x02\xfa"\xf1\x8e\xe8p\xe0\xf1\xd88\xd2g\xcc\x9a\xc7\xeb\xc3l\xc1)\xc0)\xc0l\xc1\xeb\xc3\x9a\xc7g\xcc8\xd2\xf1\xd8p\xe0\x8e\xe8"\xf1\x02\xfa\xff\x02\xee\x0b\xa1\x14\xeb\x1c\xa3$\xa2+\xc41\xeb6\xfb:\xe2=\x8f?\xfb?#?\x0c=\xc19Q5\xd3/d)$"6\x1a\xc4\x11\xf9\x08\x00\x00\x07\xf7<\xee\xca\xe5\xdc\xdd\x9c\xd6-\xd0\xaf\xca?\xc6\xf4\xc2\xdd\xc0\x05\xc0q\xc0\x1e\xc2\x05\xc5\x15\xc9<\xce^\xd4]\xdb\x15\xe3_\xeb\x12\xf4\x01\xfd\xfe\x05\xde\x0er\x17\x90\x1f\x0f\'\xc8-\x993f8\x15<\x94>\xd7?\xd7?\x94>\x15<f8\x993\xc8-\x0f\'\x90\x1fr\x17\xde\x0e\xfe\x05\x01\xfd\x12\xf4_\xeb\x15\xe3]\xdb^\xd4<\xce\x15\xc9\x05\xc5\x1e\xc2q\xc0\x05\xc0\xdd\xc0\xf4\xc2?\xc6\xaf\xca-\xd0\x9c\xd6\xdc\xdd\xca\xe5<\xee\x07\xf7'

A couple things to notice about how these clips were generated:

  1. Each of the arrays is multiplied by 2**14, so that when they are written as integers they don’t all just round to zero. In I2S, the magnitude corresponds to the volume, with \(2^15\) being the loudest possible signal (that’s the largest number you can have in a 16 bit integer, since one bit handles sign).

  2. Since we know exactly how many samples we want in each clip, we can specify that directly rather than having to compute a frequency.

  3. By using the flag endpoint=False we ensure that the signal will repeat nicely. In effect, the endpoint for one cycle is the start of the next cycle – there is no reason to include it twice.

Exercise 3 (30 pts)#

Use the sample counts you found in exercise 2 to make byte strings for the other notes. Copy them into synth.py, and enjoy your completed synthesizer!

Write Answers for Exercise 3 Below

Exercise 4 (20 pts)#

What do you think of this synthesizer so far? What are some things it does well? What are some of the limitations of this design? How would you improve on it? Answer in two paragraphs.

Write Answers for Exercise 4 Below

Advanced exercises (Optional)#

The following exercises are an opportunity to explore beyond the basic material presented above. There is not time to do all of them. Instead, pick just one of them to do for a grade. The rest can be treated as practice problems. These questions are intentionally open-ended, so you have some freedom in how you approach them. You are also welcome to work in teams of 2-3 for this exercise.

Pick the best sample count#

In order to approximate middle B, we found a rational number \(\frac{134}{3}\) which was close to the desired number of samples per cycle, \(\frac{22050}{440\times 2^{1/6}}\). The continued fraction expansion allows you to find a sequence of “best possible” rational approximations to any number, in the sense that each approximation is closer than any approximation with a smaller denominator. For example, we can use Sympy to generate the first several approximations to middle B.

import sympy
import itertools

middle_b = 22050/(440*(2**(1/sympy.S(6))))
display(sympy.Eq(middle_b, middle_b.evalf(5), evaluate=False))

best_approximations = sympy.continued_fraction_convergents(sympy.continued_fraction_iterator(middle_b))
for approx in itertools.islice(best_approximations, 6):
    display(sympy.Eq(approx, approx.evalf(5), evaluate=False))
\[\displaystyle \frac{2205 \cdot 2^{\frac{5}{6}}}{88} = 44.646\]
\[\displaystyle 44 = 44.0\]
\[\displaystyle 45 = 45.0\]
\[\displaystyle \frac{89}{2} = 44.5\]
\[\displaystyle \frac{134}{3} = 44.667\]
\[\displaystyle \frac{625}{14} = 44.643\]
\[\displaystyle \frac{759}{17} = 44.647\]

You can see that both our initial attempt (45) and our improved attempt (134/3) appear in the list.

The challenge for you is to make a software tool which does the following:

  1. Computes the best rational approximations for any note, telling the error (in cents) of each approximation.

  2. Produces a plot with error (in cents) on the y axis, and sample count on the x axis.

  3. Makes a recommendation for which approximation to use, subject to user-specified constraints.

Include different waveforms and octaves#

Most synthesizers not only let you choose the frequency of the note you play, they also allow you to select from several waveforms. Also, synthesizers with few keys will typically provide a way to shift between octaves.

Connect some additional buttons to the Pico, and write a script to have one of them cycle between modes for different waveforms and the other cycle between modes for different octaves. When it switches modes, have the Pico print the current selection on the serial console.

Polyphony#

One unfortunate property of most synthesizers is that you can’t press every key at once. That is definitely the case here, since the basic synthesizer we made only supports pressing one note at a time.

You have already designed clips for each note playing independently. Your task is to make audio clips for each pair of notes, and modify the synthesizer code so that if exactly two buttons are pressed then your two-note clip plays.