Multilayer Perceptron questions
I am working on a school project, designing a neural network (mlp),
I made it with a GUI so it can be interactive.
For all my neurons I am using SUM as GIN function,
the user can select the activation function for each layer.
I have a theoretical question:
- do I set the threshold,g and a - parameters individually for each neuron or for the entire layer?
neural-network perceptron
add a comment |
I am working on a school project, designing a neural network (mlp),
I made it with a GUI so it can be interactive.
For all my neurons I am using SUM as GIN function,
the user can select the activation function for each layer.
I have a theoretical question:
- do I set the threshold,g and a - parameters individually for each neuron or for the entire layer?
neural-network perceptron
add a comment |
I am working on a school project, designing a neural network (mlp),
I made it with a GUI so it can be interactive.
For all my neurons I am using SUM as GIN function,
the user can select the activation function for each layer.
I have a theoretical question:
- do I set the threshold,g and a - parameters individually for each neuron or for the entire layer?
neural-network perceptron
I am working on a school project, designing a neural network (mlp),
I made it with a GUI so it can be interactive.
For all my neurons I am using SUM as GIN function,
the user can select the activation function for each layer.
I have a theoretical question:
- do I set the threshold,g and a - parameters individually for each neuron or for the entire layer?
neural-network perceptron
neural-network perceptron
edited Nov 10 at 11:12
asked Nov 10 at 11:06
random_numbers
237
237
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
Looks nice ! You can have 3 hidden layers, but you'll see with experimenting, you will rarely need that many layers. What is your training pattern ?
Answer to your question depends on your training pattern and purpose of input neurons.. when e.g. some input neuron has a different type of value, you could use another threshold function or different settings for parameters in neurons connected to that input neuron.
But in general, it is better to feed neural network input into seperate perceptrons. So, the answer is: in theory, you could preset individual properties of neurons.. but in practice of back-propagation learning, it is not needed. There are no "individual properties" of neurons, the weight values that result of your training cycles will differ every time. All initial weights can be set on a small random value, transfer threshold and learning rate are to be set per layer.
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53238319%2fmultilayer-perceptron-questions%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
Looks nice ! You can have 3 hidden layers, but you'll see with experimenting, you will rarely need that many layers. What is your training pattern ?
Answer to your question depends on your training pattern and purpose of input neurons.. when e.g. some input neuron has a different type of value, you could use another threshold function or different settings for parameters in neurons connected to that input neuron.
But in general, it is better to feed neural network input into seperate perceptrons. So, the answer is: in theory, you could preset individual properties of neurons.. but in practice of back-propagation learning, it is not needed. There are no "individual properties" of neurons, the weight values that result of your training cycles will differ every time. All initial weights can be set on a small random value, transfer threshold and learning rate are to be set per layer.
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
add a comment |
Looks nice ! You can have 3 hidden layers, but you'll see with experimenting, you will rarely need that many layers. What is your training pattern ?
Answer to your question depends on your training pattern and purpose of input neurons.. when e.g. some input neuron has a different type of value, you could use another threshold function or different settings for parameters in neurons connected to that input neuron.
But in general, it is better to feed neural network input into seperate perceptrons. So, the answer is: in theory, you could preset individual properties of neurons.. but in practice of back-propagation learning, it is not needed. There are no "individual properties" of neurons, the weight values that result of your training cycles will differ every time. All initial weights can be set on a small random value, transfer threshold and learning rate are to be set per layer.
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
add a comment |
Looks nice ! You can have 3 hidden layers, but you'll see with experimenting, you will rarely need that many layers. What is your training pattern ?
Answer to your question depends on your training pattern and purpose of input neurons.. when e.g. some input neuron has a different type of value, you could use another threshold function or different settings for parameters in neurons connected to that input neuron.
But in general, it is better to feed neural network input into seperate perceptrons. So, the answer is: in theory, you could preset individual properties of neurons.. but in practice of back-propagation learning, it is not needed. There are no "individual properties" of neurons, the weight values that result of your training cycles will differ every time. All initial weights can be set on a small random value, transfer threshold and learning rate are to be set per layer.
Looks nice ! You can have 3 hidden layers, but you'll see with experimenting, you will rarely need that many layers. What is your training pattern ?
Answer to your question depends on your training pattern and purpose of input neurons.. when e.g. some input neuron has a different type of value, you could use another threshold function or different settings for parameters in neurons connected to that input neuron.
But in general, it is better to feed neural network input into seperate perceptrons. So, the answer is: in theory, you could preset individual properties of neurons.. but in practice of back-propagation learning, it is not needed. There are no "individual properties" of neurons, the weight values that result of your training cycles will differ every time. All initial weights can be set on a small random value, transfer threshold and learning rate are to be set per layer.
edited Nov 10 at 11:46
answered Nov 10 at 11:40
Goodies
41127
41127
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
add a comment |
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
We haven`t learned yet about learning patterns , ohh wait we did , supervised learning, this project is more of a GUI thing and so the teacher can see that we understand how a neural network should work, In a later project we will have to implement a network to solve a problem. And we will only be learning about feed forward, no back propagation.
– random_numbers
Nov 10 at 11:44
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
In that case I would keep it simple. User should specify the number of layers, the size of layers and a maximum random initial weight value for all weights. Use a single treshold function for all neurons.. There is one thing: you will need to set 3 learning rates in above case: when using hidden layers, user should be able to set different learning rates for each layer.
– Goodies
Nov 10 at 11:48
add a comment |
Thanks for contributing an answer to Stack Overflow!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53238319%2fmultilayer-perceptron-questions%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown