Neural network as a nonlinear system?
$begingroup$
I defined a very simple neural network of $2$ inputs, $1$ hidden layer with $2$ nodes, and one output node.
For each input pattern $x⃗ ∈ ℝ×ℝ$
and associated output $o∈ℝ$, the resulting nonlinear equation is:
$wo_{0} σ(x_0 Wi{00} + x_1 Wi{10}) + wo{1} σ(x_0 Wi{01} + x_1 Wi{11}) = o$
where $Wi$ is the weight matrix of order $2×2$, where each element $Wi_{jk} in ℝ$, of input connections, $σ(x)=frac{1}{1+exp(−x)}$, and $vec{wo}$, with $wo_{i} in ℝ$, is the weight vector of the two output connections before the output node.
Given a dataset of $n$ (pattern, output) examples, there will be $n$ nonlinear equations.
I'm asking how to find the solutions of those nonlinear systems, as an alternative method to solve the learning problem, without backpropagation. I've implemented an optimizer for the stated probem. If someone is interested I can provide the relative C sources (email: fportera2@gmail.com).
nonlinear-system
$endgroup$
add a comment |
$begingroup$
I defined a very simple neural network of $2$ inputs, $1$ hidden layer with $2$ nodes, and one output node.
For each input pattern $x⃗ ∈ ℝ×ℝ$
and associated output $o∈ℝ$, the resulting nonlinear equation is:
$wo_{0} σ(x_0 Wi{00} + x_1 Wi{10}) + wo{1} σ(x_0 Wi{01} + x_1 Wi{11}) = o$
where $Wi$ is the weight matrix of order $2×2$, where each element $Wi_{jk} in ℝ$, of input connections, $σ(x)=frac{1}{1+exp(−x)}$, and $vec{wo}$, with $wo_{i} in ℝ$, is the weight vector of the two output connections before the output node.
Given a dataset of $n$ (pattern, output) examples, there will be $n$ nonlinear equations.
I'm asking how to find the solutions of those nonlinear systems, as an alternative method to solve the learning problem, without backpropagation. I've implemented an optimizer for the stated probem. If someone is interested I can provide the relative C sources (email: fportera2@gmail.com).
nonlinear-system
$endgroup$
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28
add a comment |
$begingroup$
I defined a very simple neural network of $2$ inputs, $1$ hidden layer with $2$ nodes, and one output node.
For each input pattern $x⃗ ∈ ℝ×ℝ$
and associated output $o∈ℝ$, the resulting nonlinear equation is:
$wo_{0} σ(x_0 Wi{00} + x_1 Wi{10}) + wo{1} σ(x_0 Wi{01} + x_1 Wi{11}) = o$
where $Wi$ is the weight matrix of order $2×2$, where each element $Wi_{jk} in ℝ$, of input connections, $σ(x)=frac{1}{1+exp(−x)}$, and $vec{wo}$, with $wo_{i} in ℝ$, is the weight vector of the two output connections before the output node.
Given a dataset of $n$ (pattern, output) examples, there will be $n$ nonlinear equations.
I'm asking how to find the solutions of those nonlinear systems, as an alternative method to solve the learning problem, without backpropagation. I've implemented an optimizer for the stated probem. If someone is interested I can provide the relative C sources (email: fportera2@gmail.com).
nonlinear-system
$endgroup$
I defined a very simple neural network of $2$ inputs, $1$ hidden layer with $2$ nodes, and one output node.
For each input pattern $x⃗ ∈ ℝ×ℝ$
and associated output $o∈ℝ$, the resulting nonlinear equation is:
$wo_{0} σ(x_0 Wi{00} + x_1 Wi{10}) + wo{1} σ(x_0 Wi{01} + x_1 Wi{11}) = o$
where $Wi$ is the weight matrix of order $2×2$, where each element $Wi_{jk} in ℝ$, of input connections, $σ(x)=frac{1}{1+exp(−x)}$, and $vec{wo}$, with $wo_{i} in ℝ$, is the weight vector of the two output connections before the output node.
Given a dataset of $n$ (pattern, output) examples, there will be $n$ nonlinear equations.
I'm asking how to find the solutions of those nonlinear systems, as an alternative method to solve the learning problem, without backpropagation. I've implemented an optimizer for the stated probem. If someone is interested I can provide the relative C sources (email: fportera2@gmail.com).
nonlinear-system
nonlinear-system
edited Nov 30 '18 at 7:55
Filippo Portera
asked Nov 29 '18 at 21:27
Filippo PorteraFilippo Portera
112
112
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28
add a comment |
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019243%2fneural-network-as-a-nonlinear-system%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019243%2fneural-network-as-a-nonlinear-system%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
Genetic algorithms?
$endgroup$
– N74
Nov 29 '18 at 21:46
$begingroup$
I would like to implement a g.a. method. Are you sure it will perform better than the gradient method, if I can ask the question?
$endgroup$
– Filippo Portera
Dec 2 '18 at 18:49
$begingroup$
I am sure that, for the simple problem you are facing, the back propagation is the best way. But you asked for an alternative method, and genetic algorithms are able to span bigger parameter spaces and find global optima, instead of being locked in a local one.
$endgroup$
– N74
Dec 2 '18 at 22:28