문제

I understand a perceptron can only work correctly on linearly separable sets, like the outputs of the NAND, AND, OR functions. I've been reading Wikipedia's entry on the perceptron, and got to play with its code.

XOR is a case where a single layer perceptron should fail, as it's not a linearly separable set.

#xor
print ("xor")
t_s           = [((1, 1, 1), 0), ((1, 0, 1), 1), ((1, 1, 0), 1), ((1, 1, 1), 0)] 


threshold     = 0.5
learning_rate = 0.1
w             = [0, 0, 0]

def dot_product(values, weights):
    return sum(value * weight for value, weight in zip(values, weights))

def train_perceptron(threshold, learning_rate, weights, training_set):
    while True:
        #print('-' * 60)
        error_count = 0

        for input_vector, desired_output in training_set:
            #print(weights)
            result = dot_product(input_vector, weights) > threshold
            error  = desired_output - result

            if error != 0:
                error_count += 1
                for index, value in enumerate(input_vector):
                    weights[index] += learning_rate * error * value

        if error_count == 0: #iterate till there's no error 
            break
    return training_set

t_s = train_perceptron(threshold, learning_rate, w, t_s)

t_s = [(a[1:], b) for a, b in t_s]

for a, b in t_s:
    print "input: " + str(a) + ", output: " + str(b)

The output for this Ideone run is correct for XOR. How come?

xor
input: (1, 1), output: 0
input: (0, 1), output: 1
input: (1, 0), output: 1
input: (1, 1), output: 0
도움이 되었습니까?

해결책

You input t_s into train_perceptron and return it without modifying. Then you output it. Of course that works perfectly....

t_s = train_perceptron(threshold, learning_rate, w, t_s)

This does not change t_s at all. train_perceptron does at no point modify training_set,. but returns it: return training_set

Then here you output it:

t_s = [(a[1:], b) for a, b in t_s]

for a, b in t_s:
    print "input: " + str(a) + ", output: " + str(b)

다른 팁

Try changing your training set:

t_s = [((1, 1, 1), 0), ((1, 0, 1), 1), ((1, 1, 0), 1), ((0, 0, 0), 0)]

If my memory is correct to break non-linear problem with a perceptron you need at least one hidden layer with non-linear activation for neurons in this layer.

라이센스 : CC-BY-SA ~와 함께 속성
제휴하지 않습니다 StackOverflow
scroll top