====== Differences ====== This shows you the differences between two versions of the page.
|
deepunlearning [2013/06/30 16:15] bmcfee |
deepunlearning [2013/06/30 17:04] (current) craffel |
||
|---|---|---|---|
| Line 1: | Line 1: | ||
| - | ====== Idea ====== | + | ====== Deep Unlearning ====== |
| MIR techniques rely upon accurate representations of acoustic content in order to produce high-quality results. Over the past few decades, most research has operated on hand-crafted features, which work well up to a point, but may discard important information from the representation, thereby degrading performance. | MIR techniques rely upon accurate representations of acoustic content in order to produce high-quality results. Over the past few decades, most research has operated on hand-crafted features, which work well up to a point, but may discard important information from the representation, thereby degrading performance. | ||
| Line 14: | Line 14: | ||
| ====== Implementation ====== | ====== Implementation ====== | ||
| + | Our implementation is written in Python, using the LibROSA library for low-level audio analysis, and Theano for feature learning. | ||
| + | |||
| + | The model architecture is based upon the ''convolutional_mlp.py'' example from the DeepLearningTutorial, with the following modifications: | ||
| + | - The input layer operates on a short fragment of audio (~0.5s) represented as a $64\times 40$-dimensional Mel power spectrum. | ||
| + | - Layer 1 consists of a bank of 2-dimensional convolutional filters. Each filter is convolved with the input layer, and the resulting filter responses are downsampled by spatial max-pooling. | ||
| + | - Layer 2 consists of a linear transformation of the pooled filter responses, followed by a bank of rectified linear units | ||
| + | - Layer 3 is the output layer, which is implemented as a logistic regression classifier to predict which of $k$ known artists generated the input patch. | ||
| + | |||
| + | The model is trained by full stochastic gradient descent using a learning rate of $0.05$ and batches of 80 randomly selected input patches. The objective function is cross-entropy of the output layer against the true label, combined with $\ell_2$-regularization of the filter weights and output layer parameters. | ||
| ====== Our stuff ====== | ====== Our stuff ====== | ||
| Line 24: | Line 33: | ||
| * [[https://github.com/Theano/Theano|Theano]] | * [[https://github.com/Theano/Theano|Theano]] | ||
| * [[https://github.com/bmcfee/librosa|LibROSA]] | * [[https://github.com/bmcfee/librosa|LibROSA]] | ||
| + | |||
| + | ====== Authors ====== | ||
| + | * Brian McFee | ||
| + | * Nicola Montecchio | ||
| + | |||