source: cpp/frams/genetics/oper_fx.cpp @ 764

Last change on this file since 764 was 764, checked in by Maciej Komosinski, 6 years ago

Less deterministic (less absorbing) mutation for very narrow allowed ranges (performs reflect+wrap, previously reflect+absorb)

  • Property svn:eol-style set to native
File size: 12.1 KB
RevLine 
[286]1// This file is a part of Framsticks SDK.  http://www.framsticks.com/
[749]2// Copyright (C) 1999-2018  Maciej Komosinski and Szymon Ulatowski.
[286]3// See LICENSE.txt for details.
[109]4
5#include <ctype.h>  //isupper()
[121]6#include "oper_fx.h"
[375]7#include <common/log.h>
[109]8#include <common/nonstd_math.h>
9#include <frams/util/rndutil.h>
10
[168]11static double distrib_force[] =   // for '!'
[109]12{
[168]13        3,             // distribution 0 -__/ +1
14        0.001, 0.2,    // "slow" neurons
15        0.001, 1,
16        1, 1,          // "fast" neurons
[109]17};
[168]18static double distrib_inertia[] =  // for '='
[109]19{
[168]20        2,             // distribution 0 |..- +1
21        0, 0,          // "fast" neurons
22        0.7, 0.98,
[109]23};
[168]24static double distrib_sigmo[] =  // for '/'
[109]25{
[168]26        5,             // distribution -999 -..-^-..- +999
27        -999, -999,    //"perceptron"
28        999, 999,
29        -5, -1,        // nonlinear
30        1, 5,
31        -1, 1,         // ~linear
[109]32};
33
34
[168]35int GenoOperators::roulette(const double *probtab, const int count)
[109]36{
[168]37        double sum = 0;
38        int i;
39        for (i = 0; i < count; i++) sum += probtab[i];
40        double sel = rnd01*sum;
41        for (sum = 0, i = 0; i < count; i++) { sum += probtab[i]; if (sel < sum) return i; }
42        return -1;
[109]43}
44
[168]45bool GenoOperators::getMinMaxDef(ParamInterface *p, int i, double &mn, double &mx, double &def)
[109]46{
[168]47        mn = mx = def = 0;
48        int defined = 0;
49        if (p->type(i)[0] == 'f')
50        {
51                double _mn = 0, _mx = 1, _def = 0.5;
[743]52                defined = p->getMinMaxDouble(i, _mn, _mx, _def);
[168]53                if (defined == 1) _mx = _mn + 1.0;
54                if (_mx < _mn && defined == 3) _mn = _mx = _def; //only default was defined, let's assume min=max=default
55                if (defined < 3) _def = (_mn + _mx) / 2.0;
56                mn = _mn; mx = _mx; def = _def;
57        }
58        if (p->type(i)[0] == 'd')
59        {
[247]60                paInt _mn = 0, _mx = 1, _def = 0;
[743]61                defined = p->getMinMaxInt(i, _mn, _mx, _def);
[168]62                if (defined == 1) _mx = _mn + 1;
63                if (_mx < _mn && defined == 3) _mn = _mx = _def; //only default was defined, let's assume min=max=default
64                if (defined < 3) _def = (_mn + _mx) / 2;
65                mn = _mn; mx = _mx; def = _def;
66        }
67        return defined == 3;
[109]68}
69
[121]70int GenoOperators::selectRandomProperty(Neuro* n)
[109]71{
[168]72        int neuext = n->extraProperties().getPropCount(),
73                neucls = n->getClass() == NULL ? 0 : n->getClass()->getProperties().getPropCount();
74        if (neuext + neucls == 0) return -1; //no properties in this neuron
75        int index = randomN(neuext + neucls);
76        if (index >= neuext) index = index - neuext + 100;
77        return index;
[109]78}
79
[168]80double GenoOperators::mutateNeuProperty(double current, Neuro *n, int i)
[109]81{
[751]82        if (i == -1) return mutateCreepNoLimit('f', current, 2, true); //i==-1: mutating weight of neural connection
[168]83        Param p;
84        if (i >= 100) { i -= 100; p = n->getClass()->getProperties(); }
85        else p = n->extraProperties();
86        double newval = current;
87        /*bool ok=*/getMutatedProperty(p, i, current, newval);
88        return newval;
[109]89}
90
[168]91bool GenoOperators::mutatePropertyNaive(ParamInterface &p, int i)
[109]92{
[168]93        double mn, mx, df;
94        if (p.type(i)[0] != 'f' && p.type(i)[0] != 'd') return false; //don't know how to mutate
95        getMinMaxDef(&p, i, mn, mx, df);
[109]96
[168]97        ExtValue ev;
98        p.get(i, ev);
[751]99        ev.setDouble(mutateCreep(p.type(i)[0], ev.getDouble(), mn, mx, true));
[168]100        p.set(i, ev);
101        return true;
[109]102}
103
[168]104bool GenoOperators::mutateProperty(ParamInterface &p, int i)
[109]105{
[168]106        double newval;
107        ExtValue ev;
108        p.get(i, ev);
109        bool ok = getMutatedProperty(p, i, ev.getDouble(), newval);
110        if (ok) { ev.setDouble(newval); p.set(i, ev); }
111        return ok;
[109]112}
113
[168]114bool GenoOperators::getMutatedProperty(ParamInterface &p, int i, double oldval, double &newval)
[109]115{
[168]116        newval = 0;
117        if (p.type(i)[0] != 'f' && p.type(i)[0] != 'd') return false; //don't know how to mutate
118        const char *n = p.id(i), *na = p.name(i);
119        if (strcmp(n, "si") == 0 && strcmp(na, "Sigmoid") == 0) newval = CustomRnd(distrib_sigmo); else
120                if (strcmp(n, "in") == 0 && strcmp(na, "Inertia") == 0) newval = CustomRnd(distrib_inertia); else
121                        if (strcmp(n, "fo") == 0 && strcmp(na, "Force") == 0) newval = CustomRnd(distrib_force); else
122                        {
[670]123                double mn, mx, df;
124                getMinMaxDef(&p, i, mn, mx, df);
[751]125                newval = mutateCreep(p.type(i)[0], oldval, mn, mx, true);
[168]126                        }
127        return true;
[109]128}
129
[751]130double GenoOperators::mutateCreepNoLimit(char type, double current, double stddev, bool limit_precision_3digits)
[109]131{
[751]132        double result = RndGen.Gauss(current, stddev);
133        if (type == 'd')
134        {
135                result = int(result + 0.5);
136                if (result == current) result += randomN(2) * 2 - 1; //force some change
137        }
138        else
139        {
140                if (limit_precision_3digits)
141                        result = floor(result * 1000 + 0.5) / 1000.0; //round
142        }
[168]143        return result;
[109]144}
145
[751]146double GenoOperators::mutateCreep(char type, double current, double mn, double mx, double stddev, bool limit_precision_3digits)
[109]147{
[751]148        double result = mutateCreepNoLimit(type, current, stddev, limit_precision_3digits);
[764]149        if (result<mn || result>mx) //exceeds boundary, so bring to the allowed range
150        {
151                //reflect:
152                if (result > mx) result = mx - (result - mx); else
153                        if (result < mn) result = mn + (mn - result);
154                //wrap (just in case 'result' exceeded the allowed range so much that after reflection above it exceeded the other boundary):
155                if (result > mx) result = mn + fmod(result - mx, mx - mn); else
156                        if (result < mn) result = mn + fmod(mn - result, mx - mn);
157                if (limit_precision_3digits)
158                {
159                        //reflect and wrap above may have changed the (limited) precision, so try to round again (maybe unnecessarily, because we don't know if reflect+wrap above were triggered)
160                        double result_try = floor(result * 1000 + 0.5) / 1000.0; //round
161                        if (mn <= result_try && result_try <= mx) result = result_try; //after rounding still witin allowed range, so keep rounded value
162                }
163        }
[146]164        return result;
[109]165}
166
[751]167double GenoOperators::mutateCreep(char type, double current, double mn, double mx, bool limit_precision_3digits)
168{
169        double stddev = (mx - mn) / 2 / 5; // magic arbitrary formula for stddev, which becomes /halfinterval, 5 times narrower
170        return mutateCreep(type, current, mn, mx, stddev, limit_precision_3digits);
171}
172
[146]173void GenoOperators::setIntFromDoubleWithProbabilisticDithering(ParamInterface &p, int index, double value) //TODO
174{
[749]175        p.setInt(index, (paInt)(value + 0.5)); //TODO value=2.499 will result in 2 and 2.5 will result in 3, but we want these cases to be 2 or 3 with almost equal probability. value=2.1 should be mostly 2, rarely 3. Careful with negative values (test it!)
[146]176}
177
[749]178void GenoOperators::linearMix(vector<double> &p1, vector<double> &p2, double proportion)
179{
180        if (p1.size() != p2.size())
181        {
182                logPrintf("GenoOperators", "linearMix", LOG_ERROR, "Cannot mix vectors of different length (%d and %d)", p1.size(), p2.size());
183                return;
184        }
185        for (unsigned int i = 0; i < p1.size(); i++)
186        {
187                double v1 = p1[i];
188                double v2 = p2[i];
189                p1[i] = v1*proportion + v2*(1 - proportion);
190                p2[i] = v2*proportion + v1*(1 - proportion);
191        }
192}
193
[146]194void GenoOperators::linearMix(ParamInterface &p1, int i1, ParamInterface &p2, int i2, double proportion)
195{
[158]196        char type1 = p1.type(i1)[0];
197        char type2 = p2.type(i2)[0];
198        if (type1 == 'f' && type2 == 'f')
[146]199        {
200                double v1 = p1.getDouble(i1);
201                double v2 = p2.getDouble(i2);
202                p1.setDouble(i1, v1*proportion + v2*(1 - proportion));
203                p2.setDouble(i2, v2*proportion + v1*(1 - proportion));
204        }
[158]205        else
206                if (type1 == 'd' && type2 == 'd')
207                {
[670]208                int v1 = p1.getInt(i1);
209                int v2 = p2.getInt(i2);
210                setIntFromDoubleWithProbabilisticDithering(p1, i1, v1*proportion + v2*(1 - proportion));
211                setIntFromDoubleWithProbabilisticDithering(p2, i2, v2*proportion + v1*(1 - proportion));
[158]212                }
213                else
[375]214                        logPrintf("GenoOperators", "linearMix", LOG_WARN, "Cannot mix values of types '%c' and '%c'", type1, type2);
[146]215}
216
[121]217NeuroClass* GenoOperators::getRandomNeuroClass()
[109]218{
[673]219        vector<NeuroClass*> active;
[168]220        for (int i = 0; i < Neuro::getClassCount(); i++)
[673]221                if (Neuro::getClass(i)->genactive)
222                        active.push_back(Neuro::getClass(i));
223        if (active.size() == 0) return NULL; else return active[randomN(active.size())];
[109]224}
225
[758]226NeuroClass* GenoOperators::getRandomNeuroClassWithOutput()
227{
228        vector<NeuroClass*> active;
229        for (int i = 0; i < Neuro::getClassCount(); i++)
230                if (Neuro::getClass(i)->genactive && Neuro::getClass(i)->getPreferredOutput() != 0)
231                        active.push_back(Neuro::getClass(i));
232        if (active.size() == 0) return NULL; else return active[randomN(active.size())];
233}
234
235NeuroClass* GenoOperators::getRandomNeuroClassWithInput()
236{
237        vector<NeuroClass*> active;
238        for (int i = 0; i < Neuro::getClassCount(); i++)
239                if (Neuro::getClass(i)->genactive && Neuro::getClass(i)->getPreferredInputs() != 0)
240                        active.push_back(Neuro::getClass(i));
241        if (active.size() == 0) return NULL; else return active[randomN(active.size())];
242}
243
244NeuroClass* GenoOperators::getRandomNeuroClassWithOutputAndNoInputs()
245{
246        vector<NeuroClass*> active;
247        for (int i = 0; i < Neuro::getClassCount(); i++)
248                if (Neuro::getClass(i)->genactive && Neuro::getClass(i)->getPreferredOutput() != 0 && Neuro::getClass(i)->getPreferredInputs() == 0)
249                        active.push_back(Neuro::getClass(i));
250        if (active.size() == 0) return NULL; else return active[randomN(active.size())];
251}
252
[673]253int GenoOperators::getRandomNeuroClassWithOutput(const vector<NeuroClass*>& NClist)
254{
255        vector<int> allowed;
256        for (size_t i = 0; i < NClist.size(); i++)
257                if (NClist[i]->getPreferredOutput() != 0) //this NeuroClass provides output
258                        allowed.push_back(i);
259        if (allowed.size() == 0) return -1; else return allowed[randomN(allowed.size())];
260}
261
262int GenoOperators::getRandomNeuroClassWithInput(const vector<NeuroClass*>& NClist)
263{
264        vector<int> allowed;
265        for (size_t i = 0; i < NClist.size(); i++)
266                if (NClist[i]->getPreferredInputs() != 0) //this NeuroClass wants one input connection or more                 
267                        allowed.push_back(i);
268        if (allowed.size() == 0) return -1; else return allowed[randomN(allowed.size())];
269}
270
271int GenoOperators::getRandomChar(const char *choices, const char *excluded)
272{
273        int allowed_count = 0;
274        for (size_t i = 0; i < strlen(choices); i++) if (!strchrn0(excluded, choices[i])) allowed_count++;
275        if (allowed_count == 0) return -1; //no char is allowed
276        int rnd_index = randomN(allowed_count) + 1;
277        allowed_count = 0;
278        for (size_t i = 0; i < strlen(choices); i++)
279        {
280                if (!strchrn0(excluded, choices[i])) allowed_count++;
281                if (allowed_count == rnd_index) return i;
282        }
283        return -1; //never happens
284}
285
[121]286NeuroClass* GenoOperators::parseNeuroClass(char*& s)
[109]287{
[670]288        int maxlen = (int)strlen(s);
289        int NClen = 0;
290        NeuroClass *NC = NULL;
[168]291        for (int i = 0; i<Neuro::getClassCount(); i++)
292        {
[670]293                const char *ncname = Neuro::getClass(i)->name.c_str();
294                int ncnamelen = (int)strlen(ncname);
295                if (maxlen >= ncnamelen && ncnamelen>NClen && (strncmp(s, ncname, ncnamelen) == 0))
296                {
297                        NC = Neuro::getClass(i);
298                        NClen = ncnamelen;
299                }
[168]300        }
[670]301        s += NClen;
302        return NC;
[109]303}
304
[168]305Neuro* GenoOperators::findNeuro(const Model *m, const NeuroClass *nc)
[109]306{
[168]307        if (!m) return NULL;
308        for (int i = 0; i < m->getNeuroCount(); i++)
309                if (m->getNeuro(i)->getClass() == nc) return m->getNeuro(i);
310        return NULL; //neuron of class 'nc' was not found
[109]311}
312
[168]313int GenoOperators::neuroClassProp(char*& s, NeuroClass *nc, bool also_v1_N_props)
[109]314{
[247]315        int len = (int)strlen(s);
[168]316        int Len = 0, I = -1;
317        if (nc)
318        {
319                Param p = nc->getProperties();
320                for (int i = 0; i<p.getPropCount(); i++)
321                {
322                        const char *n = p.id(i);
[247]323                        int l = (int)strlen(n);
[168]324                        if (len >= l && l>Len && (strncmp(s, n, l) == 0)) { I = 100 + i; Len = l; }
325                        if (also_v1_N_props) //recognize old properties symbols /=!
326                        {
327                                if (strcmp(n, "si") == 0) n = "/"; else
328                                        if (strcmp(n, "in") == 0) n = "="; else
329                                                if (strcmp(n, "fo") == 0) n = "!";
[247]330                                l = (int)strlen(n);
[168]331                                if (len >= l && l > Len && (strncmp(s, n, l) == 0)) { I = 100 + i; Len = l; }
332                        }
333                }
334        }
335        Neuro n;
336        Param p = n.extraProperties();
337        for (int i = 0; i<p.getPropCount(); i++)
338        {
339                const char *n = p.id(i);
[247]340                int l = (int)strlen(n);
[168]341                if (len >= l && l>Len && (strncmp(s, n, l) == 0)) { I = i; Len = l; }
342        }
343        s += Len;
344        return I;
[109]345}
346
[121]347bool GenoOperators::isWS(const char c)
[168]348{
349        return c == ' ' || c == '\n' || c == '\t' || c == '\r';
350}
[109]351
[121]352void GenoOperators::skipWS(char *&s)
[158]353{
[168]354        if (s == NULL)
[375]355                logMessage("GenoOperators", "skipWS", LOG_WARN, "NULL reference!");
[158]356        else
[670]357                while (isWS(*s)) s++;
[109]358}
359
[168]360bool GenoOperators::areAlike(char *g1, char *g2)
[109]361{
362        while (*g1 || *g2)
363        {
364                skipWS(g1);
365                skipWS(g2);
366                if (*g1 != *g2) return false; //when difference
[168]367                if (!*g1 && !*g2) break; //both end
368                g1++;
369                g2++;
[109]370        }
371        return true; //equal
372}
373
[168]374char* GenoOperators::strchrn0(const char *str, char ch)
375{
376        return ch == 0 ? NULL : strchr((char*)str, ch);
377}
[109]378
[758]379bool GenoOperators::canStartNeuroClassName(const char firstchar)
[109]380{
[168]381        return isupper(firstchar) || firstchar == '|' || firstchar == '@' || firstchar == '*';
[109]382}
383
Note: See TracBrowser for help on using the repository browser.