Using data from the musixmatch API, I have created a tool that shows what words musical artists use most in their lyrics. It uses a combination of jQuery and D3 libraries to work as a single-page “app”. Try it here.
The web “app” works by finding the artist you search for, and running through every lyric it can find by the artist to develop a lexicon. Once that lexicon is built, you can click on the box that is created to view statistics about their most-used words.
You can also view the most-used overall by clicking the “All Artists” button.
The instant an Arduino or Raspberry Pi connects to the web (with a public IP) it is out there for anyone – or anything – to detect.
In our connected devices class, my classmates and I all saw this vulnerability firsthand. After leaving our connected thermostats on for a week, we experienced our devices being scanned and sometimes attacked by machines from across the globe.
Planet Music is a sound visualizer on the moon! The current version has a set playlist of 6 songs that can be activated with the media player buttons. The space scene reacts to the sound using a fast Fourier transform (FFT). Here is a sample video:
I’m fascinated by the way sound and images combine, and how we perceive them. I like concerts because the music is accompanied by visual compliments, whether it is lighting, dancing, or (increasingly) digital images and videos on the big screens.
In the big leagues, these visualizations are high-quality animations, often specifically designed for the song that is playing. One that always comes to mind is the animation that accompanied Tiesto’s ‘Escape Me’ during his Kaleidoscope world tour. It was many years ago, and this was the best video I could find:
Amazing production, the visuals really complimented the song; but what about music in the mid-leagues or little-leagues? Is it possible to entertain and engage people with more accessible stuff? This is the avenue I would like to explore for this final project.
My goal is to create a dynamic music visualizer – an accessible sketch that detects something in the music data and provides visual feedback in real (or near-real) time.
As the technical feasibility of the API to LED project comes together, it is time to consider the physical specifications, including design, construction, and materials. Here is the current conceptual design:
Friday night! The perfect time for expanding on the API-to-LED work from last week. The improvement now offers a more diverse set of information and utilizes duplex serial communication.
The result is that now we can now use a toggle-switch to see the 1-hour price change of either Bitcoin or Ethereum in the form of LED lights. The position of the attached toggle switch determines which currency is being shown.
The programming also features a p5 sketch that reflects the information being sent through the serial.
For example, in the video below Ethereum has had a modest positive 1 hour price change (0.06%), so it lights up the green LEDs a teeny bit. At the same time, Bitcoin has had a larger, negative price change (-1.16%), so it lights up the red LEDs quite a bit.