联系方式

  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-21:00
  • 微信:codinghelp

您当前位置:首页 >> Matlab编程Matlab编程

日期:2018-12-01 10:04

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 1/12

Homework Seven (Grads Only)

Due Dec 3 by 3pm Points 100 Submitting a file upload Available after Oct 31 at 12am

Submit Assignment

For this assignment, you will be asked to train a deep convolutional neural network from scratch.

We will be learning about many of the concepts needed in this homework assignment in the weeks of 11/12, 11/19,

and 11/26. However, you are advised to begin teaching yourself about Neural Networks *now* and refine your work as

we cover material. The MATLAB website has numerous tutorials on performing Deep Learning in

MATLAB. https://www.mathworks.com/solutions/deep-learning/examples/training-a-model-from-scratch.html

(https://www.mathworks.com/solutions/deep-learning/examples/training-a-model-from-scratch.html)

In this project, you will train a deep convolutional network from scratch to recognize scenes. The starter code gives

you a very simple network architecture which doesn't work that well and you will add jittering, normalization,

regularization, and more layers to increase recognition accuracy to 50, 60, or perhaps 70%. Unfortunately, we only

have 1,500 training examples so it doesn't seem possible to train a network from scratch which outperforms handcrafted

features.

Project materials including starter code, training and testing data, and html writeup template: hw7.zip.

You must separately download MatConvNet 1.0 beta 16. (direct link to beta 16). MAKE SURE YOU USE THE

BETA 16 VERSION.

MatConvNet isn't precompiled like VLFeat, so we compiled it for you.

Make sure you have the MATLAB Parallel Computing Toolbox installed.

Starter Code Outline

The following is an outline of the stencil code:

hw7.m . The top level function for training a deep network from scratch for scene recognition. If you run this starter

code unmodified it will train a simple network that achieves only 25% accuracy (about as good as the tiny images

baseline in project 4). hw7.m calls:

hw7_setup_data.m . Loads the 15 scene database into MatConvNet imdb format.

hw7_cnn_init.m . Initializes the convolutional network by specifying the various layers which perform convolution,

max pooling, non-linearities, normalization, regularization, and the final loss layer.

hw7_get_batch() (defined inside hw7.m ) This operates on each batch of training images to be passed into the

network during training. This is where you can "jitter" your training data.

The deep network training will be performed by cnn_train.m which in turn calls vl_simplenn.m

(http://www.vlfeat.org/matconvnet/mfiles/simplenn/vl_simplenn/) but you will not need to modify those functions for this

project.

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 2/12

Part 0

First install MatConvNet and make sure it is working, using the code below. Step through the following MatConvNet

"Quick start" demo. You can simply copy and paste the commands into the Matlab command window.

% install and compile MatConvNet

% (you can skip this if you already installed MatConvNet beta 16 and the mex files)

% untar('http://www.vlfeat.org/matconvnet/download/matconvnet-1.0-beta16.tar.gz') ;

% cd matconvnet-1.0-beta16

% run matlab/vl_compilenn

% download a pre-trained CNN from the web (needed once)

websave('imagenet-vgg-f.mat', 'http://www.vlfeat.org/matconvnet/models/imagenet-vgg-f.mat') ;

% If you have problems downloading through Matlab,

% you can simply put the url into a web browser and save the file.

% Or potentially modify the save location (and adapt below) to save somewhere where you have permissions.

% setup MatConvNet. Your path might be different.

run '../../matconvnet-1.0-beta16/matlab/vl_setupnn'

% load the 233MB pre-trained CNN

net = load('imagenet-vgg-f.mat') ;

% Fix any compatibility issues with the network

net = vl_simplenn_tidy(net) ;

% load and preprocess an image

im = imread('peppers.png') ;

im_ = single(im) ; % note: 255 range

im_ = imresize(im_, net.meta.normalization.imageSize(1:2)) ;

im_ = im_ - net.meta.normalization.averageImage ;

% run the CNN

res = vl_simplenn(net, im_) ;

% show the classification result

scores = squeeze(gather(res(end).x)) ;

[bestScore, best] = max(scores) ;

figure(1) ; clf ; imagesc(im) ;

title(sprintf('%s (%d), score %.3f',...

net.meta.classes.description{best}, best, bestScore)) ;

Troubleshooting. If you encounter errors trying to run this demo, make sure: (1) You have MatConvNet 1.0 beta 16

(not a different version). (2) Your mex files are in the correct location [MatConvNetPath]/matlab/mex/ . If you encounter

errors about invalid mex files in Windows you may be missing Visual C++ Redistributable Packages

(https://www.microsoft.com/en-us/download/details.aspx?id=40784) . If you encounter an error about

about labindex being undefined you may be missing the parallel computing toolbox for Matlab.

Before we start building our own deep convolutional networks, it might be useful to have a look at MatConvNet's

tutorial (http://www.robots.ox.ac.uk/~vgg/practicals/cnn/index.html) . In particular, you should be able to understand

Part 1 of the tutorial. In addition to the examples shown in parts 3 and 4 of the tutorial, MatConvNet has example

code (http://www.vlfeat.org/matconvnet/training/) for training networks to recognize the MNIST and CIFAR datasets.

Your project follows the same outline as those examples. Feel free to take a look at that code for inspiration. You can

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 3/12

run the example code to watch the training process MNIST and CIFAR. Training will take about 5 and 15 minutes for

those datasets, respectively.

Compiling MatConvNet with GPU support is more complex and not needed for this project. If you're trying to do extra

credit and find yourself frustrated with training times you can try, though.

Training a deep network from scratch

For the bits below, make sure you are running your code from the code directory of the hw7 directory; if you get errors,

make sure your hw7 directory and /home/tom/Downloads/matconvnet-1.0-beta16 directory are in the same higherlevel

directory.

Run hw7.m and bask in the glory of deep learning. Gone are the days of hand designed features. Now we have endto-end

learning in which a highly non-linear representation is learned for our data to maximize our objective (in this

case, 15-way classification accuracy). Instead of an anemic 70% accuracy we can now recognize scenes with... 25%

accuracy. OK, that didn't work at all. What's going on?

First, let's take a look at the network architecture used in this experiment. Here is the code from hw7_cnn_init.m that

specifies the network structure:

Let's make sure we understand what's going on here. This simple baseline network has 4 layers -- a convolutional

layer, followed by a max pool layer, followed by a rectified linear layer, followed by another convolutional layer. This

last convolutional layer might be called a "fully connected" or "fc" layer because its output has a spatial resolution of

1x1. Equivalently, every unit in the output of that layer is a function of the entire previous layer (thus "fully connected").

But mathematically there's not really any difference from "convolutional" layers so we specify them in the same way in

MatConvNet.

Let's look at the first convolutional layer. The "weights" are the filters being learned. They are initialized with random

numbers from a Gaussian distribution. The inputs to randn(9,9,1,10) mean the filters have a 9x9 spatial resolution,

span 1 filter depth (because the input images are grayscale), and that there are 10 filters. The network also learns a

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 4/12

bias or constant offset to associate with the output of each filter. This is what zeros(1,10) initializes.

The next layer is a max pooling layer. It will take a max over a 7x7 sliding window and then subsample the resulting

image / map with a stride of 7. Thus the max pooling layer will decrease the spatial resolution by a factor of 7

according to the stride parameter. The filter depth will remain the same (10). There are other pooling possibilities

(e.g. average pooling) but we will only use max pooling in this project.

The next layer is the non-linearity. Any values in the feature map from the max pooling layer which are negative will be

set to 0. There are other non-linearity possibilities (e.g. sigmoid) but we will use only rectified linear in this project.

Note that the pool layer and relu layer have no learned parameters associated with them. We are hand-specifying their

behavior entirely, so there are no weights to initialize as in the convolutional layers.

Finally, we have the last layer which is convolutional (but might be called "fully connected" because it happens to

reduce the spatial resolution to 1x1). The filters learned at this layer operate on the rectified, subsampled, maxpooled

filter responses from the first layer. The output of this layer must be 1x1 spatial resolution (or "data size") and

it must have a filter depth of 15 (corresponding to the 15 categories of the 15 scene database). This is achieved by

initializing the weights with randn(8,8,10,15) . 8x8 is the spatial resolution of the filters. 10 is the number of filter

dimensions that each of these filters take as input and 15 is the number of dimensions out. 10 is highlighted in green

to emphasize that it must be the same in those 3 places -- if the first convolutional layer has weights for 10 filters, it

must also have offsets for 10 filters, and the next convolutional layer must take as input 10 filter dimensions.

At the top of our network we add one more layer which is only used for training. This is the "loss" layer. There are

many possible loss functions but we will use the "softmax" loss for this project. This loss function will measure how

badly the network is doing for any input (i.e. how different its final layer activations are from the ground truth, where

ground truth in our case is category membership). The network weights will update, through backpropagation, based

on the derivative of the loss function. With each training batch the network weights will take a tiny gradient descent

step in the direction that should decrease the loss function (but isn't actually guaranteed to, because the steps are of

some finite length, or because dropout regularization will turn off part of the network).

How did we know to make the final layer filters have a spatial resolution of 8x8? It's not obvious because we don't

directly specify output resolution. Instead it is derived from the input image resolution and the filter widths, padding,

and strides of the previous layers. Luckily MatConvNet provides a visualization function vl_simplenn_display to help us

figure this out. Here is what it looks like if we specify the net as shown above and then call vl_simplenn_display(net,

'inputSize', [64 64 1 50]) .

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 5/12

If the last convolutional layer had a filter size of 6x6 that would lead to a "data size" in the network visualization of 3x3

and we would know we need to change things (subsample more in previous layers or create wider filters in the final

layer). In general it is not at all obvious what the right network architecture is. It takes a lot of artistry (read as: black

magic) to design the right network and training strategy for optimal performance.

We just said the network has 4 real layers but this visualization shows 6. That's because it includes a layer 0 which is

the input image and a layer 5 which is the loss layer. For each layer this visualization shows several useful attributes.

"data size" is the spatial resolution of the feature maps at each level. In this network and most deep networks this

will decrease as you move up thet network. "data depth" is the number of channels or filters in each layer. This will

tend to increase as you move up a network. "rf size" is the receptive field size. That is how large an area in the original

image a particular network unit is sensitive to. This will increase as you move up the network. Finally this visualization

shows us that the network has 10,000 free parameters, the vast majority of them associated with the last convolutional

layer.

OK, now we understand a bit about the network. Let's analyze its performance. After 30 training epochs (30 passes

through the training data) Matlab's Figure 1 should look like this:

(Note: your plots might slightly different due to updates in MATLAB.)

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 6/12

We'll be studying these figures quite a bit during this project so it's important to understand what it shows.

The left pane shows the training loss (blue) and validation loss (dashed orange) across training epochs. Each training

epoch is a pass over the entire training set of 1500 images broken up into "batches" of 50 training instances. The code

shuffles the order of the training instances randomly each epoch. When the network makes mistakes, it incurs a "loss"

and backpropagation updates the weights of the network in a direction that shoulddecrease the loss. Therefore the

blue line should more or less decrease monotonically. On the other hand, the orange line is the loss incurred on

theheld out test set. The figure refers to it as "val" or "validation". In a realistic recognition scenario we might have

three sets of data: train, validation, and test. We would use validation to assess how well our training is working and to

know when to stop training and then we would test on a completely held out test set. For this project the validation set

is our test set. We're trying to maximize performance on the validation set and that's it. The pass through the validation

set does not change the network weights in any way. The pass through the validation set is also 3 times faster than

the training pass because it does not have the "backwards" pass to update network weights.

The middle pane shows the training and testing accuracy on the train and test (val) data sets across the same training

epochs. It shows top 1 error -- how often the highest scoring guess is wrong. We're interested in top 1 error,

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 7/12

specifically the top 1 error on the held out validation / test set.

The right pane shows top 5 error -- how often all of the 5 highest scoring guesses are wrong. We're not as worried

about this metric.

In this experiment, the training and test top 1 error started out around 93% which is exactly what we would expect. If

you have 15 categories and you make a random guess on each test case, you will be wrong 93% of the time. As the

training progressed and the network weights moved away from their random initialization, accuracy increased.

Note the areas circled in green corresponding to the first 8 training epochs. During these epochs, the

training and validation error were decreasing which is exactly what we want to see. Beyond that point the error on the

training dataset kept decreasing, but the validation error did not. Our lowest error on the validation/test set is around

75% (or 25% accuracy). We are overfitting to our training data. This is hard to avoid with a small training set. In fact, if

we let this experiment run for 200 epochs we see that it is possible for the training accuracy to become perfect with no

appreciable increase in test accuracy:

Now we are going to take several steps to improve the performance of our convolutional network. The modifications

we make will familiarize you with the building blocks of deep learning that can lead to impressive accuracy with

enough training data. With the relatively small amount of training data in the 15 scene database, it is very hard to

outperform hand-crafted features.

Learning rate. Before we start making changes, there is a very important learning parameter that you might need to

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 8/12

tune any time you change the network or the data being input to the network. The learning rate (set by default

as opts.LearningRate = 0.0001 in hw7.m ) determines the size of the gradient descent steps taken by the network

weights. If things aren't working, try making it much smaller or larger (e.g. by factors of 10). If the objective remains

exactly constant over the first dozen epochs, the learning rate might have been too high and "broken" some aspect of

the network. If the objective spikes or even becomes NaN then the learning rate may also be too large. However, a

very small learning rate requires many training epochs.

Problem 1: We don't have enough training data. Let's "jitter".

If you left-right flip (mirror) an image of a scene, it never changes categories. A kitchen doesn't become a forest when

mirrored. This isn't true in all domains -- a "d" becomes a "b" when mirrored, so you can't "jitter" digit recognition

training data in the same way. But we can synthetically increase our amount of training data by left-right mirroring

training images during the learning process.

The learning process calls getBatch() in hw7.m each time it wants training or testing images. Modify getBatch() to

randomly flip some of the images (or entire batches). Useful functions: rand and fliplr .

You can try more elaborate forms of jittering -- zooming in a random amount, rotating a random amount, taking a

random crop, etc. Mirroring helps quite a bit on its own, though, and is easy to implement. You should see a 5% to

10% increase in accuracy (or drop in top 1 validation error) by adding mirroring.

After you implement mirroring, you should notice that your training error doesn't drop as quickly. That's actually a good

thing, because it means the network isn't overfitting to the 1,500 original training images as much (because it sees

3,000 training images now, although they're not as good as 3,000 truly independent samples). Because the training

and test errors fall more slowly, you may need more training epochs or you may try modifying the learning rate.

Problem 2: The images aren't zero-centered.

One simple trick which can help a lot is to subtract the mean from every image. Modify hw7_setup_data.m so that it

computes the mean image and then subtracts the mean from all images before returning imdb . It would arguably be

more proper to only compute the mean from the training images (since the test/validation images should be strictly

held out) but it won't make much of a difference. After doing this you should see another 15% or so increase in

accuracy. Most of this increase will show up in the first few iterations.

Problem 3: Our network isn't regularized.

If you train your network (especially for more than the default number of epochs) you'll see that the training error can

decrease to zero while the val top1 error hovers at 40% to 50%. The network has learned weights which can perfectly

recognize the training data, but those weights don't generalize to held out test data. The best regularization would be

more training data but we don't have that. Instead we will use dropout regularization. We add a dropout layer to our

convolutional net as follows:

What does dropout regularization do? It randomly turns off network connections at training time to fight overfitting. This

prevents a unit in one layer from relying too strongly on a single unit in the previous layer. Dropout regularization can

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 9/12

be interpreted as simultaneously training many "thinned" versions of your network. At test test, all connections are

restored which is analogous to taking an average prediction over all of the "thinned" networks. You can see a more

complete discussion of dropout regularization in this paper

(https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf) .

The dropout layer has only one free parameter -- the dropout rate -- the proportion of connections that are randomly

deleted. The default of 0.5 should be fine. Insert a dropout layer between your convolutional layers. In particular, insert

it directly before your last convolutional layer. Your test accuracy should increase by another 10%. Your train accuracy

should decrease much more slowly. That's to be expected -- you're making life much harder for the training algorithm

by cutting out connections randomly.

If you increase the number of training epochs (and maybe decrease the learning rate) you should be able to achieve

55% test accuracy (45% top1 val error) or slightly better at this point. Notice how much more structured the learned

filters are at this point compared to the initial network before we made improvements:

Problem 4: Our network isn't deep.

Let's take a moment to reflect on what our convolutional network is actually doing. We learn filters which seem to be

looking horizontal edges, vertical edges, and parallel edges. Some of the filters have diagonal orientations and some

seem to be looking for high frequencies or center-surround. This learned filter bank is applied to each input image, the

maximum response from each 7x7 block is taken by the max pooling, and then the rectified linear layer zeros out

negative values. The fully connected layer sees a 10 channel image with 8x8 spatial resolution. It learns 15 linear

classifiers (a linear filter with a learned threshold is basically a linear classifier) on this 8x8 filter response map. This

architecture is reminiscent of hand-crafted features like the gist scene descriptor

(http://people.csail.mit.edu/torralba/code/spatialenvelope/) developed precisely for scene recoginition (on 8 scene

categories which would later be incorporated into the 15 scene database). The gist descriptor actually works better

than our learned feature. The gist descriptor with a non-linear classifier can achieve 74.7% accuracy on the 15 scene

database.

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 10/12

Our convolutional network to this point isn't "deep". It has two layers with learned weights. Contrast this with the

example networks for MNIST and CIFAR in MatConvNet which contain 4 and 5 layers, respectively. AlexNet and

VGG-F contain 8 layers. The VGG "very deep" networks (http://www.robots.ox.ac.uk/~vgg/research/very_deep/)

contain 16 and 19 layers. ResNet (https://arxiv.org/abs/1512.03385) contains up to 150 layers.

One quite unsatisfying aspect of our current network architecture is that the max-pooling operation covers a window of

7x7 and then is subsampled with a stride of 7. That seems overly lossy and deep networks usually do not subsample

by more than a factor of 2 or 3 each layer.

Let's make our network deeper by adding an additional convolutional layer in hw7_cnn_init.m . In fact, we probably don't

want to add just a convolutional layer, but another max-pool layer and relu layer, as well. For example, you might

insert a convolutional layer after the existing relu layer with a 5x5 spatial support followed by a max-pool over a 3x3

window with a stride of 2. You can reduce the max-pool window in the previous layer, adjust padding, and reduce the

spatial resolution of the final layer until vl_simplenn_display(net, 'inputSize', [64 64 1 50]) , which is called at the end

of hw7_cnn_init() shows that your network's final layer (not counting the softmax) has a data size of 1 and a data

depth of 15. You also need to make sure that the data depth output by any channel matches the data depth input to

the following channel. For instance, maybe your new convolutional layer takes in the 10 channels of the first layer but

outputs 15 channels. The final layer would then need to have its weights initialized accordingly to account for the fact

that it operates on a 15 channel image instead of a 10 channel image.

Training deeper networks is tricky. The networks are slower to train and more sensitive to initialization and learning

rate. For now, try to add one or two more blocks of "conv / pool / relu" and see if you can simply match your

performance of the previous section.

Problem 5: Our "deep" network is slow to train and brittle.

You might have noticed that your deeper network doesn't seem to learn very reasonable filters in the first layer. It is

harder for the gradients to pass from the last layer all the way to the first in a deeper architecture. Normalization can

help. In particular, let's add a batch normalization (https://arxiv.org/abs/1502.03167) layer after each convolutional

layer except for the last. So if you have 4 total convolutional layers we will add 3 batch normalization layers.

Add the following code to hw7_cnn_init(). You can call this function to add batch normalization to your conv layers

after you've built your network. Be careful with the layer indexing, because calling net = insertBnorm(net,

layer_index) will add a new layer and thus shift the index of later convolutional layers.

% --------------------------------------------------------------------

function net = insertBnorm(net, l)

% --------------------------------------------------------------------

assert(isfield(net.layers{l}, 'weights'));

ndim = size(net.layers{l}.weights{1}, 4);

layer = struct('type', 'bnorm', ...

'weights', {{ones(ndim, 1, 'single'), zeros(ndim, 1, 'single')}}, ...

'learningRate', [1 1 0.05], ...

'weightDecay', [0 0]) ;

net.layers{l}.weights{2} = [] ; % eliminate bias in previous conv layer

net.layers = horzcat(net.layers(1:l), layer, net.layers(l+1:end)) ;

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 11/12

Batch normalization by itself won't necessarily increase accuracy, but it will allow you to use much higher learning

rates. Try increasing your learning rate by a factor of 10 or even 100 and now you can rapidly explore and train

different network architectures. Notice how the first layer filters start to show structure quickly with batch

normalization.

We leave it up to you to determine the specifics of your deeper network: number of layers, number of filters in each

layer, size of filters in each layer, padding, max-pooling, stride, dropout factor, etc. It is not required that your deeper

network increases accuracy over the shallow network (but it can with the right hyperparameters). As long as you can

achieve 50% test accuracy for some epoch with a deeper network which uses mirroring to jitter, zero-centers the

images as they are loaded, and regularizes the network with a dropout layer you will receive full credit. Try to keep the

total training time under 10 minutes. You can achieve high accuracy in 100 epochs and in less in less than 10

minutes.

Additional optional improvements

Enjoy chasing higher accuracy? Here's some optional directions to investigate which might help improve your

accuracy.

If you look at MatConvNet's ImageNet examples you can see that the learning rate isn't constant during training.

You can specify learning rate as pts.learningRate = logspace(-3, -5, 120) to have it change from .001 to .00001 over

120 training epochs, for instance. This form of learning rate schedule can improve performance slightly.

You can try increasing the filter depth of the network. The example networks for MNIST, CIFAR, and ImageNet

have 20, 32, and 64 filters in the first layer and it tends to increase as you go up the network.

The MNIST, CIFAR, and ImageNet examples in MatConvNet show numerous advanced strategies: Use of

normalization layers, variable learning rate per layer (the two elements of the per-layer learning rate

in cnn_cifar_init.m are the relative learning rates for the filters and the bias terms), use of average pooling instead

of max pooling for some layers, initializing weights with distributions other than randn , more dramatic jittering, etc.

The more free parameters your network has the more prone to overfitting it is. Multiple dropout layers can help

fight back against this, but will slow down training considerably.

One obvious limitation of our network is that it operates on 64x64 images when the scene images are generally

closer to 256x256. We're definitely losing valuable texture information by working at low resolution. Luckily, it's not

necessarily slow to work with the higher resolution images if you put a greater-than-one stride in your first

convolutional layer. The VGG-F network adopts this strategy. You can see in cnn_imagenet_init.m that its first layer

uses 11x11 filters with a stride of 4. This is 1/16th as many evaluations as a stride of 1.

The images can be normalized more strongly (e.g. making them have unit standard deviation) but this did not help

in my experiments.

You can try alternate loss layers at the top of your network. E.g. net.layers{end+1} = struct('name', 'hinge loss',

'type', 'loss', 'loss', 'mhinge') for hinge loss.

You can train the VGG-F network from scratch on the 15 scene database. You can call cnn_imagenet_init.m to get a

randomly initialized VGG-F and train it just like your other networks. It works better than I would expect considering

how little training data we have.

The best accuracy you can expect to achieve is 67.6% in 100 epochs, without any data augmentation beyond

mirroring and without changing the input resolution.

Write up

2018/11/22 Homework Seven (Grads Only)

https://elearning.mines.edu/courses/9621/assignments/41937 12/12

For this project, submit a PDF project report. In the report you will describe your algorithm and any decisions you

made to write your algorithm a particular way. Then you will show and discuss the results of your algorithm. We

suggest showing results plots (Matlab figure 1) and filter visualization (Matlab figure 2) as needed. In addition, submit

an archive containing all your code for this assignment. Do not hand in any networks, training data, or MatConvNet

itself! You only need to hand in 3 source files: hw7.m, hw7_cnn_init.m, and hw7_setup_data.m . You can of course hand in

any helper functions you created.

All turn-in files should be compiled into a single archive (e.g., .zip).

Credits: The design of this homework assignment comes from Georgia Tech's CV course.


版权所有:编程辅导网 2021 All Rights Reserved 联系方式:QQ:99515681 微信:codinghelp 电子信箱:99515681@qq.com
免责声明:本站部分内容从网络整理而来,只供参考!如有版权问题可联系本站删除。 站长地图

python代写
微信客服:codinghelp