Proving the learnability of XOR function by a particular neural networkhow to find the input for this...
Is this Paypal Github SDK reference really a dangerous site?
Sort array by month and year
What is Tony Stark injecting into himself in Iron Man 3?
What is the purpose of a disclaimer like "this is not legal advice"?
Will the concrete slab in a partially heated shed conduct a lot of heat to the unconditioned area?
School performs periodic password audits. Is my password compromised?
Too soon for a plot twist?
Issue with units for a rocket nozzle throat area problem
Can I challenge the interviewer to give me a proper technical feedback?
Is it appropriate to ask a former professor to order a library book for me through ILL?
Insult for someone who "doesn't know anything"
How spaceships determine each other's mass in space?
Should I file my taxes? No income, unemployed, but paid 2k in student loan interest
How do you use environments that have the same name within a single latex document?
Is divide-by-zero a security vulnerability?
Tabular environment - text vertically positions itself by bottom of tikz picture in adjacent cell
Is the differential, dp, exact or not?
If nine coins are tossed, what is the probability that the number of heads is even?
Why do phishing e-mails use faked e-mail addresses instead of the real one?
How to write a chaotic neutral protagonist and prevent my readers from thinking they are evil?
How to educate team mate to take screenshots for bugs with out unwanted stuff
Why would /etc/passwd be used every time someone executes `ls -l` command?
Should I apply for my boss's promotion?
Are small insurances worth it?
Proving the learnability of XOR function by a particular neural network
how to find the input for this optimization problem?Implement multiple input XOR with d-d-1 feed-forward neural networkWhat does the hypothesis function for a simple neural network (MLP with logistic activation, single output) look like in simplified form?Derivative of slope of neural network layer activationsApproximation rates of Neural NetworksWhat do these symbols mean regarding the stable points of a Hopfield network?Why is the Error surface for a 2 input neural network with 2 weights a parabolic bowlStudy the convexity of Mean Squared Error with regularizationIs my back propagation math correct?Are neural networks with bounded parameters a compact subset of the Banach space of continuous functions?
$begingroup$
Let's say I have the following neural network and the constraints:
- The architecture is fixed (see the network in this image, I'm not allowed to post images due to low rep) (note that there are no biases)
- Activation function for the hidden layer is $ReLU$ ;$ReLU(x) = max(0, x)$
- There's no activation function for the output layer (should just return the sum of the inputs it receive).
- Weights are constrained to be in the set ${-1, 0, 1}$
My question is:
Can we show if the XOR function is learnable or not given the network architecture and the associated constraints?
Here's how I thought about it:
Given the XOR truth table, we can right down equations for network output for each instance. If the inputs are $X_1$ and $X_2$ the output of the network $F(X_1, X_2)$ can be written as below in its general form:
$$ReLU(X_1w_1 + X_2w_3)w_5 + ReLU(X_1w_4 + X_2w_2)w_6 = F(X_1, X_2)$$
Using the truth table combinations, we obtain:
$0,1 rightarrow 1:$
$$max(0, 0 + 1.w_3)w_5 + max(0, 0 + 1w_2)w_6 = F(0, 1) = 1$$
$$max(0, w_3)w_5 + max(0, w_2)w_6 = 1 - (1)$$
$1,0 rightarrow 1:$
$$max(0, 1.w_1 + 0)w_5 + max(0, 1w_4 + 0)w_6 = F(1, 0) = 1$$
$$max(0, w_1)w_5 + max(0, w_4)w_6 = 1 - (2)$$
$1,1 rightarrow 0:$
$$max(0, 1.w_1 + 1.w_3)w_5 + max(0, 1w_4 + 1.w_2)w_6 = F(1, 1) = 0$$
$$max(0, w_1+w_3)w_5 + max(0, w_4+w_2)w_6 = 0 - (3)$$
Can we show that the above system of equations do/do not have a solution for $w_i$ values?
Here is a similar problem on crossvalidated.
proof-writing systems-of-equations constraints neural-networks constraint-programming
New contributor
$endgroup$
add a comment |
$begingroup$
Let's say I have the following neural network and the constraints:
- The architecture is fixed (see the network in this image, I'm not allowed to post images due to low rep) (note that there are no biases)
- Activation function for the hidden layer is $ReLU$ ;$ReLU(x) = max(0, x)$
- There's no activation function for the output layer (should just return the sum of the inputs it receive).
- Weights are constrained to be in the set ${-1, 0, 1}$
My question is:
Can we show if the XOR function is learnable or not given the network architecture and the associated constraints?
Here's how I thought about it:
Given the XOR truth table, we can right down equations for network output for each instance. If the inputs are $X_1$ and $X_2$ the output of the network $F(X_1, X_2)$ can be written as below in its general form:
$$ReLU(X_1w_1 + X_2w_3)w_5 + ReLU(X_1w_4 + X_2w_2)w_6 = F(X_1, X_2)$$
Using the truth table combinations, we obtain:
$0,1 rightarrow 1:$
$$max(0, 0 + 1.w_3)w_5 + max(0, 0 + 1w_2)w_6 = F(0, 1) = 1$$
$$max(0, w_3)w_5 + max(0, w_2)w_6 = 1 - (1)$$
$1,0 rightarrow 1:$
$$max(0, 1.w_1 + 0)w_5 + max(0, 1w_4 + 0)w_6 = F(1, 0) = 1$$
$$max(0, w_1)w_5 + max(0, w_4)w_6 = 1 - (2)$$
$1,1 rightarrow 0:$
$$max(0, 1.w_1 + 1.w_3)w_5 + max(0, 1w_4 + 1.w_2)w_6 = F(1, 1) = 0$$
$$max(0, w_1+w_3)w_5 + max(0, w_4+w_2)w_6 = 0 - (3)$$
Can we show that the above system of equations do/do not have a solution for $w_i$ values?
Here is a similar problem on crossvalidated.
proof-writing systems-of-equations constraints neural-networks constraint-programming
New contributor
$endgroup$
add a comment |
$begingroup$
Let's say I have the following neural network and the constraints:
- The architecture is fixed (see the network in this image, I'm not allowed to post images due to low rep) (note that there are no biases)
- Activation function for the hidden layer is $ReLU$ ;$ReLU(x) = max(0, x)$
- There's no activation function for the output layer (should just return the sum of the inputs it receive).
- Weights are constrained to be in the set ${-1, 0, 1}$
My question is:
Can we show if the XOR function is learnable or not given the network architecture and the associated constraints?
Here's how I thought about it:
Given the XOR truth table, we can right down equations for network output for each instance. If the inputs are $X_1$ and $X_2$ the output of the network $F(X_1, X_2)$ can be written as below in its general form:
$$ReLU(X_1w_1 + X_2w_3)w_5 + ReLU(X_1w_4 + X_2w_2)w_6 = F(X_1, X_2)$$
Using the truth table combinations, we obtain:
$0,1 rightarrow 1:$
$$max(0, 0 + 1.w_3)w_5 + max(0, 0 + 1w_2)w_6 = F(0, 1) = 1$$
$$max(0, w_3)w_5 + max(0, w_2)w_6 = 1 - (1)$$
$1,0 rightarrow 1:$
$$max(0, 1.w_1 + 0)w_5 + max(0, 1w_4 + 0)w_6 = F(1, 0) = 1$$
$$max(0, w_1)w_5 + max(0, w_4)w_6 = 1 - (2)$$
$1,1 rightarrow 0:$
$$max(0, 1.w_1 + 1.w_3)w_5 + max(0, 1w_4 + 1.w_2)w_6 = F(1, 1) = 0$$
$$max(0, w_1+w_3)w_5 + max(0, w_4+w_2)w_6 = 0 - (3)$$
Can we show that the above system of equations do/do not have a solution for $w_i$ values?
Here is a similar problem on crossvalidated.
proof-writing systems-of-equations constraints neural-networks constraint-programming
New contributor
$endgroup$
Let's say I have the following neural network and the constraints:
- The architecture is fixed (see the network in this image, I'm not allowed to post images due to low rep) (note that there are no biases)
- Activation function for the hidden layer is $ReLU$ ;$ReLU(x) = max(0, x)$
- There's no activation function for the output layer (should just return the sum of the inputs it receive).
- Weights are constrained to be in the set ${-1, 0, 1}$
My question is:
Can we show if the XOR function is learnable or not given the network architecture and the associated constraints?
Here's how I thought about it:
Given the XOR truth table, we can right down equations for network output for each instance. If the inputs are $X_1$ and $X_2$ the output of the network $F(X_1, X_2)$ can be written as below in its general form:
$$ReLU(X_1w_1 + X_2w_3)w_5 + ReLU(X_1w_4 + X_2w_2)w_6 = F(X_1, X_2)$$
Using the truth table combinations, we obtain:
$0,1 rightarrow 1:$
$$max(0, 0 + 1.w_3)w_5 + max(0, 0 + 1w_2)w_6 = F(0, 1) = 1$$
$$max(0, w_3)w_5 + max(0, w_2)w_6 = 1 - (1)$$
$1,0 rightarrow 1:$
$$max(0, 1.w_1 + 0)w_5 + max(0, 1w_4 + 0)w_6 = F(1, 0) = 1$$
$$max(0, w_1)w_5 + max(0, w_4)w_6 = 1 - (2)$$
$1,1 rightarrow 0:$
$$max(0, 1.w_1 + 1.w_3)w_5 + max(0, 1w_4 + 1.w_2)w_6 = F(1, 1) = 0$$
$$max(0, w_1+w_3)w_5 + max(0, w_4+w_2)w_6 = 0 - (3)$$
Can we show that the above system of equations do/do not have a solution for $w_i$ values?
Here is a similar problem on crossvalidated.
proof-writing systems-of-equations constraints neural-networks constraint-programming
proof-writing systems-of-equations constraints neural-networks constraint-programming
New contributor
New contributor
New contributor
asked yesterday
j.Doej.Doe
1
1
New contributor
New contributor
add a comment |
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
j.Doe is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3139356%2fproving-the-learnability-of-xor-function-by-a-particular-neural-network%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
j.Doe is a new contributor. Be nice, and check out our Code of Conduct.
j.Doe is a new contributor. Be nice, and check out our Code of Conduct.
j.Doe is a new contributor. Be nice, and check out our Code of Conduct.
j.Doe is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3139356%2fproving-the-learnability-of-xor-function-by-a-particular-neural-network%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown