| import math |
| import random |
| from dreamcoder.utilities import * |
|
|
|
|
| class InvalidLoss(Exception): |
| pass |
|
|
|
|
| class DN(object): |
| '''differentiable node: parent object of every differentiable operation''' |
|
|
| def __init__(self, arguments): |
| self.gradient = None |
| if arguments != []: |
| self.data = None |
| self.arguments = arguments |
|
|
| |
| |
| |
| self.descendents = [] |
|
|
| self.recalculate() |
|
|
| def __str__(self): |
| if self.arguments == []: |
| return self.name |
| return "(%s %s)" % (self.name, " ".join(str(x) |
| for x in self.arguments)) |
|
|
| def __repr__(self): |
| return "DN(op = %s, data = %s, grad = %s, #descendents = %d, args = %s)" % ( |
| self.name, self.data, self.gradient, len(self.descendents), self.arguments) |
|
|
| @property |
| def derivative(self): return self.differentiate() |
|
|
| def differentiate(self): |
| if self.gradient is None: |
| self.gradient = sum(partial * descendent.differentiate() |
| for descendent, partial in self.descendents) |
| return self.gradient |
|
|
| def zeroEverything(self): |
| if self.gradient is None and self.descendents == [] and ( |
| self.data is None or self.arguments == []): |
| return |
|
|
| self.gradient = None |
| self.descendents = [] |
| if self.arguments != []: |
| self.data = None |
|
|
| for x in self.arguments: |
| x.zeroEverything() |
|
|
| def lightweightRecalculate(self): |
| return self.forward(*[a.lightweightRecalculate() |
| for a in self.arguments]) |
|
|
| def recalculate(self): |
| if self.data is None: |
| inputs = [a.recalculate() for a in self.arguments] |
| self.data = self.forward(*inputs) |
| |
| |
| |
| |
| |
| |
| |
| partials = self.backward(*inputs) |
| for d, a in zip(partials, self.arguments): |
| |
| |
| |
| |
| |
| a.descendents.append((self, d)) |
| return self.data |
|
|
| def backPropagation(self): |
| self.gradient = 1. |
| self.recursivelyDifferentiate() |
|
|
| def recursivelyDifferentiate(self): |
| self.differentiate() |
| for x in self.arguments: |
| x.recursivelyDifferentiate() |
|
|
| def updateNetwork(self): |
| self.zeroEverything() |
| l = self.recalculate() |
| self.backPropagation() |
| return l |
|
|
| def log(self): return Logarithm(self) |
|
|
| def square(self): return Square(self) |
|
|
| def exp(self): return Exponentiation(self) |
|
|
| def clamp(self, l, u): return Clamp(self, l, u) |
|
|
| def __abs__(self): return AbsoluteValue(self) |
|
|
| def __add__(self, o): return Addition(self, Placeholder.maybe(o)) |
|
|
| def __radd__(self, o): return Addition(self, Placeholder.maybe(o)) |
|
|
| def __sub__(self, o): return Subtraction(self, Placeholder.maybe(o)) |
|
|
| def __rsub__(self, o): return Subtraction(Placeholder.maybe(o), self) |
|
|
| def __mul__(self, o): return Multiplication(self, Placeholder.maybe(o)) |
|
|
| def __rmul__(self, o): return Multiplication(self, Placeholder.maybe(o)) |
|
|
| def __neg__(self): return Negation(self) |
|
|
| def __truediv__(self, o): return Division(self, Placeholder.maybe(o)) |
|
|
| def __rtruediv__(self, o): return Division(Placeholder.maybe(o), self) |
|
|
| def numericallyVerifyGradients(self, parameters): |
| calculatedGradients = [p.derivative for p in parameters] |
| e = 0.00001 |
| for j, p in enumerate(parameters): |
| p.data -= e |
| y1 = self.lightweightRecalculate() |
| p.data += 2 * e |
| y2 = self.lightweightRecalculate() |
| p.data -= e |
| d = (y2 - y1) / (2 * e) |
| if abs(calculatedGradients[j] - d) > 0.1: |
| eprint( |
| "Bad gradient: expected %f, got %f" % |
| (d, calculatedGradients[j])) |
|
|
| def gradientDescent( |
| self, |
| parameters, |
| _=None, |
| lr=0.001, |
| steps=10**3, |
| update=None): |
| for j in range(steps): |
| l = self.updateNetwork() |
| if update is not None and j % update == 0: |
| eprint("LOSS:", l) |
| for p in parameters: |
| eprint(p.data, '\t', p.derivative) |
| if invalid(l): |
| raise InvalidLoss() |
|
|
| for p in parameters: |
| p.data -= lr * p.derivative |
| return self.data |
|
|
| def restartingOptimize(self, parameters, _=None, attempts=1, |
| s=1., decay=0.5, grow=0.1, |
| lr=0.1, steps=10**3, update=None): |
| ls = [] |
| for _ in range(attempts): |
| for p in parameters: |
| p.data = random.random()*10 - 5 |
| ls.append( |
| self.resilientBackPropagation( |
| parameters, lr=lr, steps=steps, |
| decay=decay, grow=grow)) |
| return min(ls) |
|
|
| def resilientBackPropagation( |
| self, |
| parameters, |
| _=None, |
| decay=0.5, |
| grow=1.2, |
| lr=0.1, |
| steps=10**3, |
| update=None): |
| previousSign = [None] * len(parameters) |
| lr = [lr] * len(parameters) |
| for j in range(steps): |
| l = self.updateNetwork() |
|
|
| if update is not None and j % update == 0: |
| eprint("LOSS:", l) |
| eprint("\t".join(str(p.derivative) for p in parameters)) |
| if invalid(l): |
| raise InvalidLoss() |
|
|
| newSigns = [p.derivative > 0 for p in parameters] |
| for i, p in enumerate(parameters): |
| if p.derivative > 0: |
| p.data -= lr[i] |
| elif p.derivative < 0: |
| p.data += lr[i] |
| if previousSign[i] is not None: |
| if previousSign[i] == newSigns[i]: |
| lr[i] *= grow |
| else: |
| lr[i] *= decay |
| previousSign = newSigns |
|
|
| return self.data |
|
|
|
|
| class Placeholder(DN): |
| COUNTER = 0 |
|
|
| def __init__(self, initialValue=0., name=None): |
| self.data = initialValue |
| super(Placeholder, self).__init__([]) |
| if name is None: |
| name = "p_" + str(Placeholder.COUNTER) |
| Placeholder.COUNTER += 1 |
| self.name = name |
|
|
| @staticmethod |
| def named(namePrefix, initialValue=0.): |
| p = Placeholder(initialValue, namePrefix + str(Placeholder.COUNTER)) |
| Placeholder.COUNTER += 1 |
| return p |
|
|
| def __str__(self): |
| return "Placeholder(%s = %s)" % (self.name, self.data) |
|
|
| @staticmethod |
| def maybe(x): |
| if isinstance(x, DN): |
| return x |
| return Placeholder(float(x)) |
|
|
| def forward(self): return self.data |
|
|
| def backward(self): return [] |
|
|
|
|
| class Clamp(DN): |
| def __init__(self, x, l, u): |
| assert u > l |
| self.l = l |
| self.u = u |
| super(Clamp, self).__init__([x]) |
| self.name = "clamp" |
|
|
| def forward(self, x): |
| if x > self.u: |
| return self.u |
| if x < self.l: |
| return self.l |
| return x |
|
|
| def backward(self, x): |
| if x > self.u or x < self.l: |
| return [0.] |
| else: |
| return [1.] |
|
|
|
|
| class Addition(DN): |
| def __init__(self, x, y): |
| super(Addition, self).__init__([x, y]) |
| self.name = '+' |
|
|
| def forward(self, x, y): return x + y |
|
|
| def backward(self, x, y): return [1., 1.] |
|
|
|
|
| class Subtraction(DN): |
| def __init__(self, x, y): |
| super(Subtraction, self).__init__([x, y]) |
| self.name = '-' |
|
|
| def forward(self, x, y): return x - y |
|
|
| def backward(self, x, y): return [1., -1.] |
|
|
|
|
| class Negation(DN): |
| def __init__(self, x): |
| super(Negation, self).__init__([x]) |
| self.name = '-' |
|
|
| def forward(self, x): return -x |
|
|
| def backward(self, x): return [-1.] |
|
|
|
|
| class AbsoluteValue(DN): |
| def __init__(self, x): |
| super(AbsoluteValue, self).__init__([x]) |
| self.name = 'abs' |
|
|
| def forward(self, x): return abs(x) |
|
|
| def backward(self, x): |
| if x > 0: |
| return [1.] |
| return [-1.] |
|
|
|
|
| class Multiplication(DN): |
| def __init__(self, x, y): |
| super(Multiplication, self).__init__([x, y]) |
| self.name = '*' |
|
|
| def forward(self, x, y): return x * y |
|
|
| def backward(self, x, y): return [y, x] |
|
|
|
|
| class Division(DN): |
| def __init__(self, x, y): |
| super(Division, self).__init__([x, y]) |
| self.name = '/' |
|
|
| def forward(self, x, y): return x / y |
|
|
| def backward(self, x, y): return [1.0 / y, -x / (y * y)] |
|
|
|
|
| class Square(DN): |
| def __init__(self, x): |
| super(Square, self).__init__([x]) |
| self.name = 'sq' |
|
|
| def forward(self, x): return x * x |
|
|
| def backward(self, x): return [2 * x] |
|
|
|
|
| class Exponentiation(DN): |
| def __init__(self, x): |
| super(Exponentiation, self).__init__([x]) |
| self.name = 'exp' |
|
|
| def forward(self, x): return math.exp(x) |
|
|
| def backward(self, x): return [math.exp(x)] |
|
|
|
|
| class Logarithm(DN): |
| def __init__(self, x): |
| super(Logarithm, self).__init__([x]) |
| self.name = 'log' |
|
|
| def forward(self, x): return math.log(x) |
|
|
| def backward(self, x): return [1. / x] |
|
|
|
|
| class LSE(DN): |
| def __init__(self, xs): |
| super(LSE, self).__init__(xs) |
| self.name = 'LSE' |
|
|
| def forward(self, *xs): |
| m = max(xs) |
| return m + math.log(sum(math.exp(y - m) for y in xs)) |
|
|
| def backward(self, *xs): |
| m = max(xs) |
| zm = sum(math.exp(x - m) for x in xs) |
| return [math.exp(x - m) / zm for x in xs] |
|
|
|
|
| if __name__ == "__main__": |
| x = Placeholder(10., "x") |
| y = Placeholder(2., "y") |
| z = x - LSE([x, y]) |
| z.updateNetwork() |
| eprint("dL/dx = %f\tdL/dy = %f" % (x.derivative, y.derivative)) |
|
|
| x.data = 2. |
| y.data = 10. |
| z.updateNetwork() |
| eprint("dL/dx = %f\tdL/dy = %f" % (x.differentiate(), y.differentiate())) |
|
|
| x.data = 2. |
| y.data = 2. |
| z.updateNetwork() |
| eprint("z = ", z.data, z) |
| eprint("dL/dx = %f\tdL/dy = %f" % (x.differentiate(), y.differentiate())) |
|
|
| loss = -z |
| eprint(loss) |
|
|
| lr = 0.001 |
| loss.gradientDescent([x, y], steps=10000, update=1000) |
|
|