I'm a Music Technology Researcher, Jazz Pianist, and Music Producer. My research involves using signal processing and machine learning techniques for the generation and analysis of music and other audio signals. This includes research areas like Music Information Retrieval, Machine Listening, Computational Musicology, and more. I do my work as a PhD student in the Music and Audio Research Lab (MARL) as a Steinhardt Fellow under the supervision of Dr. Juan P. Bello and Dr. Lelia Adu-Gilmore. For many years, I have been a dedicated student in both the arts and STEM, with degrees in Music Technology and Computer Engineering. My work aims to bridge these two schools of thought to create something unique.
Sound is rich with information about the surrounding environment. If you stand on a city sidewalk with your eyes closed and listen, you will hear the sounds of events happening around you: birds chirping, squirrels scurrying, people talking, doors opening, an ambulance speeding, a truck idling. In addition, you will also likely be able to perceive the location of each sound source, where it’s going, and how fast it’s moving. This project will build innovative technologies to allow computers to extract this rich information out of sound.Read More
Automatic synthesizer programming is the task of transforming an audio signal that was generated from a virtual instrument, into the parameters of a sound synthesizer that would generate this signal. In the past, this could only be done for one virtual instrument. The goal of this project is to explore techniques for automatic synthesizer programming for multi virtual instruments.
AFRINUM aims at renewing the discourse about innovation in Western African societies through the lens of the popular music industry. We consider digital technology as a globalized tool vs a cultural artifact to reflect upon the many specificities that this tool takes and the values to which it refers. This digital music culture will be investigated within a large range of social, economic and political contexts to which this culture participates meanwhile contributes to transforming, producing or reifying.Read More
Synth 3D is a VST sound synthesizer that allows users to place virtual oscillator objects in 3D space. This is accomplished by convolving oscillator signals with Head Related Impulse Responses (HRIRs) from the Spacial Audio for Domestic Interactive Entertainment Project (SADIE). The synthesizer comes with standard subtractive synthesis features. This includes three oscillators which are all individually convolved with an HRIR from SADIE, two filters, and ADSR envelopes for the filters and amplifier. Synth 3D is available as a VST3 synthesizer and Audio Units component for both Windows and Mac. The software was implemented using the JUCE framework and all of the code is open source and available to view on Github.Read More
Developed by Dr. Brian McFee, Librosa is a python package for music and audio analysis. It provides the building blocks necessary to create music information retrieval systems. Version 0.9 has expanded the library's API to provide seamless processing of multi channel audio signals.Read More
This is an optimization research project focussed on novel approaches to linear program solvers. It was done in the University of Nebraska Omaha's Data and Decisions Sciences Laboratory under the supervision of Dr. Fabio Torres Vitor. The project focuses on adding multidimensional searches into the traditional interior point method algorithm.Read More
Research on wearble computing done with the Embedded & Pervaisve Systems Lab at Washington State University under the direction of Dr. Hassan Ghasemzadeh. The goal was to use biometric signal data to predict the likelood of alchohol relapse. Participants wore Empatica's E4 wristband while performing regular tasks. The signal data was analyzed for correlation with alchohol cravings and binary classification models were developed to predict the likelood of giving in to these cravings.Read More
ProstatID™ is a software interface for post-processing MRI using cloud based computing, physician workstation integration, and AI image interpretation to provide highly accurate prostate cancer diagnostics.Read More
The Wi-Fi Synthesizer project was conceptualized as a device to read midi input and generate audio waveforms based on the information provided by the midi signal. The device would also transfer the midi signal received via Wi-Fi to a website by a http request. The midi component of the project was not included in the final application because of time restraints, but the circuitry was still included in the design. The final application outputs a 100hz audio signal with a user selected wave shape (selected from four options). The device will send data to a website indicating which waveform was last selected.Read More
The objective of this project was to give students real world experience with building a microprocessor system. The microprocessor system is based on Intel's 8051 microcontroller and performs a variety of memory functions to read, edit, move, and find contents in ram. It also interfaces with a variety of IO devices such as an LCD display, a real time clock, a seven segment display, and an analog to digital converter.Read More