Compare commits

..

No commits in common. "master" and "sj_connect_reducer_ui" have entirely different histories.

78 changed files with 623 additions and 6148 deletions

21
LICENSE
View file

@ -1,21 +0,0 @@
MIT License
Copyright (c) 2021 Sorrel
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.

110
README.md
View file

@ -1,111 +1,3 @@
# Feature Change Applier # Phono Change Applier
[Try the app!](https://sorrelbri.github.io/feature-change-applier/)
[Inspired by the Zompist Sound Change Applier 2](https://www.zompist.com/sca2.html) [Inspired by the Zompist Sound Change Applier 2](https://www.zompist.com/sca2.html)
## What is this?
Feature Change Applier is a tool for applying systemic sound change rules to an input lexicon.
Features:
- feature based phone definitions
- feature based sound change rule support
- multi-character phone support
- comparative runs for multiple rule sets
## What is LATL?
[Read the specification](/src/utils/latl/README.md)
LATL is a JavaScript targeting compiled language for doing linguistic analysis and transformations.
## How do I use FCA?
An FCA run requires the user to define three parameters:
- [the input lexicon](#The-Input-Lexicon), expressed in phonetic terms
- [the feature set](#the-feature-set), which maps each phonetic feature to positive or negative values for each phone
- [at least one 'epoch' of sound change rules](#epochs) to apply to the input lexicon
### The Input Lexicon
For best effect, the input lexicon should use a narrow phonetic transcription of each lexeme.
Lexemes must be separated by line breaks in order to be parsed properly by FCA.
Multi-word lexemes can be inserted with or without spaces, any white-space will be removed from the lexeme at runtime.
FCA does not currently support suprasegmentals by default, however features can be used to define prosodic information so long as it can be associated with a single phone.
For example:
- For tonemes, use IPA tone markers as in `ma˨˩˦` (马)
- For phonetic length, use IPA length markers as in `ħaːsin` (حَاسِن‎)
- For stress or syllabic breaks, however `ˈɡʊd.nɪs` may result in unpredictable behavior and is best avoided.
See below for defining these features in the feature set.
#### Future Changes to the Input Lexicon
- Future versions of FCA will allow for greater suprasegmental feature support.
- Future versions will allow for epoch specific lexemes
### The Feature Set
Phones in FCA are defined by the features they exhibit.
To add or edit a feature use the form to enter the feature name and the phones which are associated with the feature in the `+` and `-` inputs.
Phones should be separated by a forward slash and may be represented with multiple characters.
For example:
`aspirated + tʰ / pʰ / kʰ - t / p / k / ʔ`
Results in:
`[+ aspirated] = tʰ / pʰ / kʰ [- aspirated] = t / p / k / ʔ`
If the feature already exists, any phones associated with that feature will be replaced with the phones in the form.
A feature is not required to have a value for every phone, and every phone is not required to have a value for every feature.
Rules targeting `-` values for specific features will not target phones that are not defined in the feature set.
For example:
`[- aspirated]>ʔ/[+ vowel]_.`
This rule will not operate on the phone `ʊ` in `haʊs` as it was not defined with a negative `aspirated` value above.
#### Suprasegmentals
Toneme example using Mandarin tone system:
```
[+ tone] = ˥ / ˧˥ / ˨˩˦ / ˥˩ [- tone] =
[+ high] = ˥ / ˥˩ [- high] = ˧˥ / ˨˩˦
[+ low] = ˨˩˦ [- low] = ˥ / ˥˩ / ˧˥
[+ rising] = ˧˥ [- rising] = ˥ / ˨˩˦ / ˥˩
[+ falling] = ˨˩˦ / ˥˩ [- falling] = ˥ / ˧˥
```
Length example using Modern Standard Arabic (without allophonic variation):
```
[+ long] = aː / iː / uː [- long] = a / i / u / aj / aw
[+ geminate] = mː / nː / tː / tˤː / ... [- geminate] = m / n / t / tˤ / ...
```
#### Future Chagnes to the Feature Set
- Future versions of FCA will allow for greater suprasegmental feature support
- Future versions will allow for exclusive features. In the example below a phone cannot have a labial value and a coronal value simultaneously:
```
[!place
[labial
[+ labiodental] = f
[- labiodental] = p / m / kp / ŋm
[+ labiovelar] = kp / ŋm
[- labiovelar] = f / p / m
]
[coronal
[+ anterior] = t̪ / n̪ / t / n
[- anterior] = c / ɲ / ʈ / ɳ
[+ distributed] = t̪ / n̪ / c / ɲ
[- distributed] = t / n / ʈ / ɳ
]
...
]
```
### Epochs
This is where the rules to transform your lexicon are defined.
An FCA project can have as many 'epochs' or suites of sound change rules as you would like to define.
Rules can be defined using phones or features:
- `n>ŋ/._k`
- `[+ nasal alveolar]>[- alveolar + velar]/._[+ velar]`
These two rules will both act on the phone `n` in the sequence `nk` transforming it into `ŋ`, however the feature defined rule could also transform the `n` in the sequences `ng`, `nŋ`, `nx`, etc.
By default, FCA will pipe the initial lexicon into each one of these epochs and apply their transformations independently.
The output of one epoch can be piped into another epoch, however, by defining the `parent` parameter from the dropdown when adding a new epoch.

View file

261
package-lock.json generated
View file

@ -1,5 +1,5 @@
{ {
"name": "feature-change-applier", "name": "phono-change-applier",
"version": "0.1.0", "version": "0.1.0",
"lockfileVersion": 1, "lockfileVersion": 1,
"requires": true, "requires": true,
@ -1032,9 +1032,9 @@
"integrity": "sha512-1dVNHT76Uu5N3eJNTYcvxee+jzX4Z9lfciqRRHCU27ihbUcYi+iSc2iml5Ke1LXe1SyJCLA0+14Jh4tXJgOppA==" "integrity": "sha512-1dVNHT76Uu5N3eJNTYcvxee+jzX4Z9lfciqRRHCU27ihbUcYi+iSc2iml5Ke1LXe1SyJCLA0+14Jh4tXJgOppA=="
}, },
"@hapi/hoek": { "@hapi/hoek": {
"version": "8.5.1", "version": "8.5.0",
"resolved": "https://registry.npmjs.org/@hapi/hoek/-/hoek-8.5.1.tgz", "resolved": "https://registry.npmjs.org/@hapi/hoek/-/hoek-8.5.0.tgz",
"integrity": "sha512-yN7kbciD87WzLGc5539Tn0sApjyiGHAJgKvG9W8C7O+6c7qmoQMfVs0W4bX17eqz6C78QJqqFrtgdK5EWf6Qow==" "integrity": "sha512-7XYT10CZfPsH7j9F1Jmg1+d0ezOux2oM2GfArAzLwWe4mE2Dr3hVjsAL6+TFY49RRJlCdJDMw3nJsLFroTc8Kw=="
}, },
"@hapi/joi": { "@hapi/joi": {
"version": "15.1.1", "version": "15.1.1",
@ -4857,11 +4857,6 @@
} }
} }
}, },
"discontinuous-range": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/discontinuous-range/-/discontinuous-range-1.0.0.tgz",
"integrity": "sha1-44Mx8IRLukm5qctxx3FYWqsbxlo="
},
"dns-equal": { "dns-equal": {
"version": "1.0.0", "version": "1.0.0",
"resolved": "https://registry.npmjs.org/dns-equal/-/dns-equal-1.0.0.tgz", "resolved": "https://registry.npmjs.org/dns-equal/-/dns-equal-1.0.0.tgz",
@ -5018,11 +5013,6 @@
"minimalistic-crypto-utils": "^1.0.0" "minimalistic-crypto-utils": "^1.0.0"
} }
}, },
"email-addresses": {
"version": "3.1.0",
"resolved": "https://registry.npmjs.org/email-addresses/-/email-addresses-3.1.0.tgz",
"integrity": "sha512-k0/r7GrWVL32kZlGwfPNgB2Y/mMXVTq/decgLczm/j34whdaspNrZO8CnXPf1laaHxI6ptUlsnAxN+UAPw+fzg=="
},
"emoji-regex": { "emoji-regex": {
"version": "8.0.0", "version": "8.0.0",
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz",
@ -6106,30 +6096,6 @@
"integrity": "sha512-0Zt+s3L7Vf1biwWZ29aARiVYLx7iMGnEUl9x33fbB/j3jR81u/O2LbqK+Bm1CDSNDKVtJ/YjwY7TUd5SkeLQLw==", "integrity": "sha512-0Zt+s3L7Vf1biwWZ29aARiVYLx7iMGnEUl9x33fbB/j3jR81u/O2LbqK+Bm1CDSNDKVtJ/YjwY7TUd5SkeLQLw==",
"optional": true "optional": true
}, },
"filename-reserved-regex": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/filename-reserved-regex/-/filename-reserved-regex-1.0.0.tgz",
"integrity": "sha1-5hz4BfDeHJhFZ9A4bcXfUO5a9+Q="
},
"filenamify": {
"version": "1.2.1",
"resolved": "https://registry.npmjs.org/filenamify/-/filenamify-1.2.1.tgz",
"integrity": "sha1-qfL/0RxQO+0wABUCknI3jx8TZaU=",
"requires": {
"filename-reserved-regex": "^1.0.0",
"strip-outer": "^1.0.0",
"trim-repeated": "^1.0.0"
}
},
"filenamify-url": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/filenamify-url/-/filenamify-url-1.0.0.tgz",
"integrity": "sha1-syvYExnvWGO3MHi+1Q9GpPeXX1A=",
"requires": {
"filenamify": "^1.0.0",
"humanize-url": "^1.0.0"
}
},
"filesize": { "filesize": {
"version": "3.6.1", "version": "3.6.1",
"resolved": "https://registry.npmjs.org/filesize/-/filesize-3.6.1.tgz", "resolved": "https://registry.npmjs.org/filesize/-/filesize-3.6.1.tgz",
@ -6484,33 +6450,6 @@
"assert-plus": "^1.0.0" "assert-plus": "^1.0.0"
} }
}, },
"gh-pages": {
"version": "2.2.0",
"resolved": "https://registry.npmjs.org/gh-pages/-/gh-pages-2.2.0.tgz",
"integrity": "sha512-c+yPkNOPMFGNisYg9r4qvsMIjVYikJv7ImFOhPIVPt0+AcRUamZ7zkGRLHz7FKB0xrlZ+ddSOJsZv9XAFVXLmA==",
"requires": {
"async": "^2.6.1",
"commander": "^2.18.0",
"email-addresses": "^3.0.1",
"filenamify-url": "^1.0.0",
"fs-extra": "^8.1.0",
"globby": "^6.1.0"
},
"dependencies": {
"globby": {
"version": "6.1.0",
"resolved": "https://registry.npmjs.org/globby/-/globby-6.1.0.tgz",
"integrity": "sha1-9abXDoOV4hyFj7BInWTfAkJNUGw=",
"requires": {
"array-union": "^1.0.1",
"glob": "^7.0.3",
"object-assign": "^4.0.1",
"pify": "^2.0.0",
"pinkie-promise": "^2.0.0"
}
}
}
},
"glob": { "glob": {
"version": "7.1.6", "version": "7.1.6",
"resolved": "https://registry.npmjs.org/glob/-/glob-7.1.6.tgz", "resolved": "https://registry.npmjs.org/glob/-/glob-7.1.6.tgz",
@ -6599,12 +6538,12 @@
} }
}, },
"globule": { "globule": {
"version": "1.3.1", "version": "1.2.1",
"resolved": "https://registry.npmjs.org/globule/-/globule-1.3.1.tgz", "resolved": "https://registry.npmjs.org/globule/-/globule-1.2.1.tgz",
"integrity": "sha512-OVyWOHgw29yosRHCHo7NncwR1hW5ew0W/UrvtwvjefVJeQ26q4/8r8FmPsSF1hJ93IgWkyv16pCTz6WblMzm/g==", "integrity": "sha512-g7QtgWF4uYSL5/dn71WxubOrS7JVGCnFPEnoeChJmBnyR9Mw8nGoEwOgJL/RC2Te0WhbsEUCejfH8SZNJ+adYQ==",
"requires": { "requires": {
"glob": "~7.1.1", "glob": "~7.1.1",
"lodash": "~4.17.12", "lodash": "~4.17.10",
"minimatch": "~3.0.2" "minimatch": "~3.0.2"
} }
}, },
@ -6618,11 +6557,6 @@
"resolved": "https://registry.npmjs.org/growly/-/growly-1.3.0.tgz", "resolved": "https://registry.npmjs.org/growly/-/growly-1.3.0.tgz",
"integrity": "sha1-8QdIy+dq+WS3yWyTxrzCivEgwIE=" "integrity": "sha1-8QdIy+dq+WS3yWyTxrzCivEgwIE="
}, },
"gud": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/gud/-/gud-1.0.0.tgz",
"integrity": "sha512-zGEOVKFM5sVPPrYs7J5/hYEw2Pof8KCyOwyhG8sAF26mCAeUFAcYPu1mwB7hhpIP29zOIBaDqwuHdLp0jvZXjw=="
},
"gzip-size": { "gzip-size": {
"version": "5.1.1", "version": "5.1.1",
"resolved": "https://registry.npmjs.org/gzip-size/-/gzip-size-5.1.1.tgz", "resolved": "https://registry.npmjs.org/gzip-size/-/gzip-size-5.1.1.tgz",
@ -6769,19 +6703,6 @@
"resolved": "https://registry.npmjs.org/hex-color-regex/-/hex-color-regex-1.1.0.tgz", "resolved": "https://registry.npmjs.org/hex-color-regex/-/hex-color-regex-1.1.0.tgz",
"integrity": "sha512-l9sfDFsuqtOqKDsQdqrMRk0U85RZc0RtOR9yPI7mRVOa4FsR/BVnZ0shmQRM96Ji99kYZP/7hn1cedc1+ApsTQ==" "integrity": "sha512-l9sfDFsuqtOqKDsQdqrMRk0U85RZc0RtOR9yPI7mRVOa4FsR/BVnZ0shmQRM96Ji99kYZP/7hn1cedc1+ApsTQ=="
}, },
"history": {
"version": "4.10.1",
"resolved": "https://registry.npmjs.org/history/-/history-4.10.1.tgz",
"integrity": "sha512-36nwAD620w12kuzPAsyINPWJqlNbij+hpK1k9XRloDtym8mxzGYl2c17LnV6IAGB2Dmg4tEa7G7DlawS0+qjew==",
"requires": {
"@babel/runtime": "^7.1.2",
"loose-envify": "^1.2.0",
"resolve-pathname": "^3.0.0",
"tiny-invariant": "^1.0.2",
"tiny-warning": "^1.0.0",
"value-equal": "^1.0.1"
}
},
"hmac-drbg": { "hmac-drbg": {
"version": "1.0.1", "version": "1.0.1",
"resolved": "https://registry.npmjs.org/hmac-drbg/-/hmac-drbg-1.0.1.tgz", "resolved": "https://registry.npmjs.org/hmac-drbg/-/hmac-drbg-1.0.1.tgz",
@ -6792,14 +6713,6 @@
"minimalistic-crypto-utils": "^1.0.1" "minimalistic-crypto-utils": "^1.0.1"
} }
}, },
"hoist-non-react-statics": {
"version": "3.3.2",
"resolved": "https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-3.3.2.tgz",
"integrity": "sha512-/gGivxi8JPKWNm/W0jSmzcMPpfpPLc3dY/6GxhX2hQ9iGj3aDfklV4ET7NjKpSinLpJ5vafa9iiGIEZg10SfBw==",
"requires": {
"react-is": "^16.7.0"
}
},
"hosted-git-info": { "hosted-git-info": {
"version": "2.8.5", "version": "2.8.5",
"resolved": "https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.5.tgz", "resolved": "https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.5.tgz",
@ -6973,15 +6886,6 @@
"resolved": "https://registry.npmjs.org/https-browserify/-/https-browserify-1.0.0.tgz", "resolved": "https://registry.npmjs.org/https-browserify/-/https-browserify-1.0.0.tgz",
"integrity": "sha1-7AbBDgo0wPL68Zn3/X/Hj//QPHM=" "integrity": "sha1-7AbBDgo0wPL68Zn3/X/Hj//QPHM="
}, },
"humanize-url": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/humanize-url/-/humanize-url-1.0.1.tgz",
"integrity": "sha1-9KuZ4NKIF0yk4eUEB8VfuuRk7/8=",
"requires": {
"normalize-url": "^1.0.0",
"strip-url-auth": "^1.0.0"
}
},
"iconv-lite": { "iconv-lite": {
"version": "0.4.24", "version": "0.4.24",
"resolved": "https://registry.npmjs.org/iconv-lite/-/iconv-lite-0.4.24.tgz", "resolved": "https://registry.npmjs.org/iconv-lite/-/iconv-lite-0.4.24.tgz",
@ -7345,9 +7249,12 @@
"integrity": "sha1-qIwCU1eR8C7TfHahueqXc8gz+MI=" "integrity": "sha1-qIwCU1eR8C7TfHahueqXc8gz+MI="
}, },
"is-finite": { "is-finite": {
"version": "1.1.0", "version": "1.0.2",
"resolved": "https://registry.npmjs.org/is-finite/-/is-finite-1.1.0.tgz", "resolved": "https://registry.npmjs.org/is-finite/-/is-finite-1.0.2.tgz",
"integrity": "sha512-cdyMtqX/BOqqNBBiKlIVkytNHm49MtMlYyn1zxzvJKWmFMlGzm+ry5BBfYyeY9YmNKbRSo/o7OX9w9ale0wg3w==" "integrity": "sha1-zGZ3aVYCvlUO8R6LSqYwU0K20Ko=",
"requires": {
"number-is-nan": "^1.0.0"
}
}, },
"is-fullwidth-code-point": { "is-fullwidth-code-point": {
"version": "1.0.0", "version": "1.0.0",
@ -9256,9 +9163,9 @@
} }
}, },
"js-base64": { "js-base64": {
"version": "2.5.2", "version": "2.5.1",
"resolved": "https://registry.npmjs.org/js-base64/-/js-base64-2.5.2.tgz", "resolved": "https://registry.npmjs.org/js-base64/-/js-base64-2.5.1.tgz",
"integrity": "sha512-Vg8czh0Q7sFBSUMWWArX/miJeBWYBPpdU/3M/DKSaekLMqrqVPaedp+5mZhie/r0lgrcaYBfwXatEew6gwgiQQ==" "integrity": "sha512-M7kLczedRMYX4L8Mdh4MzyAMM9O5osx+4FcOQuTvr3A9F2D9S5JXheN0ewNbrvK2UatkTRhL5ejGmGSjNMiZuw=="
}, },
"js-levenshtein": { "js-levenshtein": {
"version": "1.1.6", "version": "1.1.6",
@ -9853,16 +9760,6 @@
"integrity": "sha1-z8RcN+nsDY8KDsPdTvf3w6vjklY=", "integrity": "sha1-z8RcN+nsDY8KDsPdTvf3w6vjklY=",
"dev": true "dev": true
}, },
"mini-create-react-context": {
"version": "0.3.2",
"resolved": "https://registry.npmjs.org/mini-create-react-context/-/mini-create-react-context-0.3.2.tgz",
"integrity": "sha512-2v+OeetEyliMt5VHMXsBhABoJ0/M4RCe7fatd/fBy6SMiKazUSEt3gxxypfnk2SHMkdBYvorHRoQxuGoiwbzAw==",
"requires": {
"@babel/runtime": "^7.4.0",
"gud": "^1.0.0",
"tiny-warning": "^1.0.2"
}
},
"mini-css-extract-plugin": { "mini-css-extract-plugin": {
"version": "0.8.0", "version": "0.8.0",
"resolved": "https://registry.npmjs.org/mini-css-extract-plugin/-/mini-css-extract-plugin-0.8.0.tgz", "resolved": "https://registry.npmjs.org/mini-css-extract-plugin/-/mini-css-extract-plugin-0.8.0.tgz",
@ -10015,11 +9912,6 @@
} }
} }
}, },
"moo": {
"version": "0.5.1",
"resolved": "https://registry.npmjs.org/moo/-/moo-0.5.1.tgz",
"integrity": "sha512-I1mnb5xn4fO80BH9BLcF0yLypy2UKl+Cb01Fu0hJRkJjlCRtxZMWkTdAtDd5ZqCOxtCkhmRwyI57vWT+1iZ67w=="
},
"move-concurrently": { "move-concurrently": {
"version": "1.0.1", "version": "1.0.1",
"resolved": "https://registry.npmjs.org/move-concurrently/-/move-concurrently-1.0.1.tgz", "resolved": "https://registry.npmjs.org/move-concurrently/-/move-concurrently-1.0.1.tgz",
@ -10092,18 +9984,6 @@
"resolved": "https://registry.npmjs.org/natural-compare/-/natural-compare-1.4.0.tgz", "resolved": "https://registry.npmjs.org/natural-compare/-/natural-compare-1.4.0.tgz",
"integrity": "sha1-Sr6/7tdUHywnrPspvbvRXI1bpPc=" "integrity": "sha1-Sr6/7tdUHywnrPspvbvRXI1bpPc="
}, },
"nearley": {
"version": "2.19.1",
"resolved": "https://registry.npmjs.org/nearley/-/nearley-2.19.1.tgz",
"integrity": "sha512-xq47GIUGXxU9vQg7g/y1o1xuKnkO7ev4nRWqftmQrLkfnE/FjRqDaGOUakM8XHPn/6pW3bGjU2wgoJyId90rqg==",
"requires": {
"commander": "^2.19.0",
"moo": "^0.5.0",
"railroad-diagrams": "^1.0.0",
"randexp": "0.4.6",
"semver": "^5.4.1"
}
},
"negotiator": { "negotiator": {
"version": "0.6.2", "version": "0.6.2",
"resolved": "https://registry.npmjs.org/negotiator/-/negotiator-0.6.2.tgz", "resolved": "https://registry.npmjs.org/negotiator/-/negotiator-0.6.2.tgz",
@ -10251,9 +10131,9 @@
} }
}, },
"node-sass": { "node-sass": {
"version": "4.13.1", "version": "4.13.0",
"resolved": "https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz", "resolved": "https://registry.npmjs.org/node-sass/-/node-sass-4.13.0.tgz",
"integrity": "sha512-TTWFx+ZhyDx1Biiez2nB0L3YrCZ/8oHagaDalbuBSlqXgUPsdkUSzJsVxeDO9LtPB49+Fh3WQl3slABo6AotNw==", "integrity": "sha512-W1XBrvoJ1dy7VsvTAS5q1V45lREbTlZQqFbiHb3R3OTTCma0XBtuG6xZ6Z4506nR4lmHPTqVRwxT6KgtWC97CA==",
"requires": { "requires": {
"async-foreach": "^0.1.3", "async-foreach": "^0.1.3",
"chalk": "^1.1.1", "chalk": "^1.1.1",
@ -12191,20 +12071,6 @@
"performance-now": "^2.1.0" "performance-now": "^2.1.0"
} }
}, },
"railroad-diagrams": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/railroad-diagrams/-/railroad-diagrams-1.0.0.tgz",
"integrity": "sha1-635iZ1SN3t+4mcG5Dlc3RVnN234="
},
"randexp": {
"version": "0.4.6",
"resolved": "https://registry.npmjs.org/randexp/-/randexp-0.4.6.tgz",
"integrity": "sha512-80WNmd9DA0tmZrw9qQa62GPPWfuXJknrmVmLcxvq4uZBdYqb1wYoKTmnlGUchvVWe0XiLupYkBoXVOxz3C8DYQ==",
"requires": {
"discontinuous-range": "1.0.0",
"ret": "~0.1.10"
}
},
"randombytes": { "randombytes": {
"version": "2.1.0", "version": "2.1.0",
"resolved": "https://registry.npmjs.org/randombytes/-/randombytes-2.1.0.tgz", "resolved": "https://registry.npmjs.org/randombytes/-/randombytes-2.1.0.tgz",
@ -12490,52 +12356,6 @@
"resolved": "https://registry.npmjs.org/react-is/-/react-is-16.12.0.tgz", "resolved": "https://registry.npmjs.org/react-is/-/react-is-16.12.0.tgz",
"integrity": "sha512-rPCkf/mWBtKc97aLL9/txD8DZdemK0vkA3JMLShjlJB3Pj3s+lpf1KaBzMfQrAmhMQB0n1cU/SUGgKKBCe837Q==" "integrity": "sha512-rPCkf/mWBtKc97aLL9/txD8DZdemK0vkA3JMLShjlJB3Pj3s+lpf1KaBzMfQrAmhMQB0n1cU/SUGgKKBCe837Q=="
}, },
"react-router": {
"version": "5.1.2",
"resolved": "https://registry.npmjs.org/react-router/-/react-router-5.1.2.tgz",
"integrity": "sha512-yjEuMFy1ONK246B+rsa0cUam5OeAQ8pyclRDgpxuSCrAlJ1qN9uZ5IgyKC7gQg0w8OM50NXHEegPh/ks9YuR2A==",
"requires": {
"@babel/runtime": "^7.1.2",
"history": "^4.9.0",
"hoist-non-react-statics": "^3.1.0",
"loose-envify": "^1.3.1",
"mini-create-react-context": "^0.3.0",
"path-to-regexp": "^1.7.0",
"prop-types": "^15.6.2",
"react-is": "^16.6.0",
"tiny-invariant": "^1.0.2",
"tiny-warning": "^1.0.0"
},
"dependencies": {
"isarray": {
"version": "0.0.1",
"resolved": "https://registry.npmjs.org/isarray/-/isarray-0.0.1.tgz",
"integrity": "sha1-ihis/Kmo9Bd+Cav8YDiTmwXR7t8="
},
"path-to-regexp": {
"version": "1.8.0",
"resolved": "https://registry.npmjs.org/path-to-regexp/-/path-to-regexp-1.8.0.tgz",
"integrity": "sha512-n43JRhlUKUAlibEJhPeir1ncUID16QnEjNpwzNdO3Lm4ywrBpBZ5oLD0I6br9evr1Y9JTqwRtAh7JLoOzAQdVA==",
"requires": {
"isarray": "0.0.1"
}
}
}
},
"react-router-dom": {
"version": "5.1.2",
"resolved": "https://registry.npmjs.org/react-router-dom/-/react-router-dom-5.1.2.tgz",
"integrity": "sha512-7BPHAaIwWpZS074UKaw1FjVdZBSVWEk8IuDXdB+OkLb8vd/WRQIpA4ag9WQk61aEfQs47wHyjWUoUGGZxpQXew==",
"requires": {
"@babel/runtime": "^7.1.2",
"history": "^4.9.0",
"loose-envify": "^1.3.1",
"prop-types": "^15.6.2",
"react-router": "5.1.2",
"tiny-invariant": "^1.0.2",
"tiny-warning": "^1.0.0"
}
},
"react-scripts": { "react-scripts": {
"version": "3.3.0", "version": "3.3.0",
"resolved": "https://registry.npmjs.org/react-scripts/-/react-scripts-3.3.0.tgz", "resolved": "https://registry.npmjs.org/react-scripts/-/react-scripts-3.3.0.tgz",
@ -12951,11 +12771,6 @@
"resolved": "https://registry.npmjs.org/resolve-from/-/resolve-from-3.0.0.tgz", "resolved": "https://registry.npmjs.org/resolve-from/-/resolve-from-3.0.0.tgz",
"integrity": "sha1-six699nWiBvItuZTM17rywoYh0g=" "integrity": "sha1-six699nWiBvItuZTM17rywoYh0g="
}, },
"resolve-pathname": {
"version": "3.0.0",
"resolved": "https://registry.npmjs.org/resolve-pathname/-/resolve-pathname-3.0.0.tgz",
"integrity": "sha512-C7rARubxI8bXFNB/hqcp/4iUeIXJhJZvFPFPiSPRnhU5UPxzMFIl+2E6yY6c4k9giDJAhtV+enfA+G89N6Csng=="
},
"resolve-url": { "resolve-url": {
"version": "0.2.1", "version": "0.2.1",
"resolved": "https://registry.npmjs.org/resolve-url/-/resolve-url-0.2.1.tgz", "resolved": "https://registry.npmjs.org/resolve-url/-/resolve-url-0.2.1.tgz",
@ -14051,19 +13866,6 @@
"resolved": "https://registry.npmjs.org/strip-json-comments/-/strip-json-comments-3.0.1.tgz", "resolved": "https://registry.npmjs.org/strip-json-comments/-/strip-json-comments-3.0.1.tgz",
"integrity": "sha512-VTyMAUfdm047mwKl+u79WIdrZxtFtn+nBxHeb844XBQ9uMNTuTHdx2hc5RiAJYqwTj3wc/xe5HLSdJSkJ+WfZw==" "integrity": "sha512-VTyMAUfdm047mwKl+u79WIdrZxtFtn+nBxHeb844XBQ9uMNTuTHdx2hc5RiAJYqwTj3wc/xe5HLSdJSkJ+WfZw=="
}, },
"strip-outer": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/strip-outer/-/strip-outer-1.0.1.tgz",
"integrity": "sha512-k55yxKHwaXnpYGsOzg4Vl8+tDrWylxDEpknGjhTiZB8dFRU5rTo9CAzeycivxV3s+zlTKwrs6WxMxR95n26kwg==",
"requires": {
"escape-string-regexp": "^1.0.2"
}
},
"strip-url-auth": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/strip-url-auth/-/strip-url-auth-1.0.1.tgz",
"integrity": "sha1-IrD6OkE4WzO+PzMVUbu4N/oM164="
},
"style-loader": { "style-loader": {
"version": "1.0.0", "version": "1.0.0",
"resolved": "https://registry.npmjs.org/style-loader/-/style-loader-1.0.0.tgz", "resolved": "https://registry.npmjs.org/style-loader/-/style-loader-1.0.0.tgz",
@ -14444,16 +14246,6 @@
"resolved": "https://registry.npmjs.org/timsort/-/timsort-0.3.0.tgz", "resolved": "https://registry.npmjs.org/timsort/-/timsort-0.3.0.tgz",
"integrity": "sha1-QFQRqOfmM5/mTbmiNN4R3DHgK9Q=" "integrity": "sha1-QFQRqOfmM5/mTbmiNN4R3DHgK9Q="
}, },
"tiny-invariant": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/tiny-invariant/-/tiny-invariant-1.1.0.tgz",
"integrity": "sha512-ytxQvrb1cPc9WBEI/HSeYYoGD0kWnGEOR8RY6KomWLBVhqz0RgTwVO9dLrGz7dC+nN9llyI7OKAgRq8Vq4ZBSw=="
},
"tiny-warning": {
"version": "1.0.3",
"resolved": "https://registry.npmjs.org/tiny-warning/-/tiny-warning-1.0.3.tgz",
"integrity": "sha512-lBN9zLN/oAf68o3zNXYrdCt1kP8WsiGW8Oo2ka41b2IM5JL/S1CTyX1rW0mb/zSuJun0ZUrDxx4sqvYS2FWzPA=="
},
"tmp": { "tmp": {
"version": "0.0.33", "version": "0.0.33",
"resolved": "https://registry.npmjs.org/tmp/-/tmp-0.0.33.tgz", "resolved": "https://registry.npmjs.org/tmp/-/tmp-0.0.33.tgz",
@ -14539,14 +14331,6 @@
"resolved": "https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz", "resolved": "https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz",
"integrity": "sha1-WIeWa7WCpFA6QetST301ARgVphM=" "integrity": "sha1-WIeWa7WCpFA6QetST301ARgVphM="
}, },
"trim-repeated": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/trim-repeated/-/trim-repeated-1.0.0.tgz",
"integrity": "sha1-42RqLqTokTEr9+rObPsFOAvAHCE=",
"requires": {
"escape-string-regexp": "^1.0.2"
}
},
"true-case-path": { "true-case-path": {
"version": "1.0.3", "version": "1.0.3",
"resolved": "https://registry.npmjs.org/true-case-path/-/true-case-path-1.0.3.tgz", "resolved": "https://registry.npmjs.org/true-case-path/-/true-case-path-1.0.3.tgz",
@ -14877,11 +14661,6 @@
"spdx-expression-parse": "^3.0.0" "spdx-expression-parse": "^3.0.0"
} }
}, },
"value-equal": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/value-equal/-/value-equal-1.0.1.tgz",
"integrity": "sha512-NOJ6JZCAWr0zlxZt+xqCHNTEKOsrks2HQd4MqhP1qy4z1SkbEP467eNx6TgDKXMvUOb+OENfJCZwM+16n7fRfw=="
},
"vary": { "vary": {
"version": "1.1.2", "version": "1.1.2",
"resolved": "https://registry.npmjs.org/vary/-/vary-1.1.2.tgz", "resolved": "https://registry.npmjs.org/vary/-/vary-1.1.2.tgz",

View file

@ -1,30 +1,21 @@
{ {
"name": "feature-change-applier", "name": "phono-change-applier",
"version": "0.1.0", "version": "0.1.0",
"private": true, "private": true,
"homepage": "https://sorrelbri.github.io/feature-change-applier",
"dependencies": { "dependencies": {
"flow-bin": "^0.113.0", "flow-bin": "^0.113.0",
"gh-pages": "^2.2.0",
"local-storage": "^2.0.0", "local-storage": "^2.0.0",
"moo": "^0.5.1", "node-sass": "^4.13.0",
"nearley": "^2.19.1",
"node-sass": "^4.13.1",
"react": "^16.12.0", "react": "^16.12.0",
"react-dom": "^16.12.0", "react-dom": "^16.12.0",
"react-router-dom": "^5.1.2",
"react-scripts": "^3.3.0" "react-scripts": "^3.3.0"
}, },
"scripts": { "scripts": {
"start": "react-scripts start", "start": "react-scripts start",
"compile-grammar": "nearleyc src/utils/latl/grammar.ne -o src/utils/latl/grammar.js",
"test-grammar": "nearley-test src/utils/latl/grammar.js --input",
"flow": "flow", "flow": "flow",
"build": "react-scripts build", "build": "react-scripts build",
"test": "react-scripts test", "test": "react-scripts test",
"eject": "react-scripts eject", "eject": "react-scripts eject"
"predeploy": "npm run build",
"deploy": "gh-pages -d build"
}, },
"eslintConfig": { "eslintConfig": {
"extends": "react-app" "extends": "react-app"

Binary file not shown.

Before

Width:  |  Height:  |  Size: 90 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 152 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 153 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 148 KiB

Binary file not shown.

Binary file not shown.

Binary file not shown.

Before

Width:  |  Height:  |  Size: 318 B

After

Width:  |  Height:  |  Size: 22 KiB

View file

@ -5,9 +5,7 @@
<link rel="icon" href="%PUBLIC_URL%/favicon.ico" /> <link rel="icon" href="%PUBLIC_URL%/favicon.ico" />
<meta name="viewport" content="width=device-width, initial-scale=1" /> <meta name="viewport" content="width=device-width, initial-scale=1" />
<link rel="manifest" href="%PUBLIC_URL%/manifest.json" /> <link rel="manifest" href="%PUBLIC_URL%/manifest.json" />
<link rel="stylesheet" href="%PUBLIC_URL%/stylesheets/reset.css"> <title>Phono Change Applier</title>
<link href="https://fonts.googleapis.com/css?family=Catamaran|Fira+Code&display=swap" rel="stylesheet">
<title>Feature Change Applier</title>
</head> </head>
<body> <body>
<noscript>You need to enable JavaScript to run this app.</noscript> <noscript>You need to enable JavaScript to run this app.</noscript>

View file

@ -1,88 +0,0 @@
set NASAL_PULMONIC_CONSONANTS = [ m̥, m, ɱ, n̼, n̥, n, ɳ̊, ɳ, ɲ̊, ɲ, ŋ, ̊ŋ, ɴ ],
STOP_PULMONIC_CONSONANTS = [ p, b, p̪, b̪, t̼, d̼, t, d, ʈ, ɖ, c, ɟ, k, ɡ, q, ɢ, ʡ, ʔ ],
S_FRICATIVE_PULMONIC_CONSONANTS = [ s, z, ʃ, ʒ, ʂ, ʐ, ɕ, ʑ ],
FRICATIVE_PULMONIC_CONSONANTS = [ ɸ, β, f, v, θ̼, ð̼, θ, ð, θ̠, ð̠, ɹ̠̊˔, ɹ̠˔, ɻ˔, ç, ʝ, x, ɣ, χ, ʁ, ħ, ʕ, h, ɦ ],
APPROXIMANT_PULMONIC_CONSONANTS = [ ʋ̥, ʋ, ɹ̥, ɹ, ɻ̊, ɻ, j̊, j, ɰ̊, ɰ, ʔ̞ ],
TAP_PULMONIC_CONSONANTS = [ ⱱ̟, ⱱ, ɾ̼, ɾ̥, ɾ, ɽ̊, ɽ, ɢ̆, ʡ̆ ],
TRILL_PULMONIC_CONSONANTS = [ ʙ̥, ʙ, r̥, r, ɽ̊r̥, ɽr, ʀ̥, ʀ, ʜ, ʢ ],
L_FRICATIVE_PULMONIC_CONSONANTS = [ ɬ, ɮ, ɭ̊˔, ɭ˔, ʎ̝̊, ʎ̝, ʟ̝̊, ʟ̝ ],
L_APPROXIMANT_PULMONIC_CONSONANTS = [ l̥, l, ɭ̊, ɭ, ʎ̥, ʎ, ʟ̥, ʟ, ʟ̠ ],
L_TAP_PULMONIC_CONSONANTS = [ ɺ, ɭ̆, ʎ̆, ʟ̆ ],
AFFRICATE_PULMONIC_CONSONANTS = [ pɸ, bβ, p̪f, b̪v, t̪θ, d̪ð, tɹ̝̊, dɹ̝, t̠ɹ̠̊˔, d̠ɹ̠˔, cç, ɟʝ, kx, ɡɣ, qχ, ʡʢ, ʔh ],
S_AFFRICATE_PULMONIC_CONSONANTS = [ ts, dz, t̠ʃ, d̠ʒ, ʈʂ, ɖʐ, tɕ, dʑ ],
L_AFFRICATE_PULMONIC_CONSONANTS = [ tɬ, dɮ, ʈɭ̊˔, cʎ̝̊, kʟ̝̊, ɡʟ̝ ],
DOUBLE_STOP_PULMONIC_CONSONANTS = [ t͡p, d͡b, k͡p, ɡ͡b, q͡ʡ ],
DOUBLE_NASAL_PULMONIC_CONSONANTS = [ n͡m, ŋ͡m ],
DOUBLE_FRICATIVE_PULMONIC_CONSONANTS = [ ɧ ],
DOUBLE_APPROXIMANT_PULMONIC_CONSONANTS = [ ʍ, w, ɥ̊, ɥ, ɫ ]
set PULMONIC_CONSONANTS, C = { NASAL_PULMONIC_CONSONANTS or STOP_PULMONIC_CONSONANTS
or S_FRICATIVE_PULMONIC_CONSONANTS or FRICATIVE_PULMONIC_CONSONANTS
or APPROXIMANT_PULMONIC_CONSONANTS or TAP_PULMONIC_CONSONANTS
or TRILL_PULMONIC_CONSONANTS or L_FRICATIVE_PULMONIC_CONSONANTS
or L_APPROXIMANT_PULMONIC_CONSONANTS or L_TAP_PULMONIC_CONSONANTS
or AFFRICATE_PULMONIC_CONSONANTS or S_AFFRICATE_PULMONIC_CONSONANTS
or L_AFFRICATE_PULMONIC_CONSONANTS or DOUBLE_STOP_PULMONIC_CONSONANTS
or DOUBLE_NASAL_PULMONIC_CONSONANTS or DOUBLE_FRICATIVE_PULMONIC_CONSONANTS
or DOUBLE_APPROXIMANT_PULMONIC_CONSONANTS
}
set STOP_EJECTIVE_CONSONANTS = [ pʼ, tʼ, ʈʼ, cʼ, kʼ, qʼ, ʡʼ ],
FRICATIVE_EJECTIVE_CONSONANTS = [ ɸʼ, fʼ, θʼ, sʼ, ʃʼ, ʂʼ, ɕʼ, xʼ, χʼ ],
L_FRICATIVE_EJECTIVE_CONSONANTS = [ ɬʼ ],
AFFRICATE_EJECTIVE_CONSONANTS = [ tsʼ, t̠ʃʼ, ʈʂʼ, kxʼ, qχʼ ],
L_AFFRICATE_EJECTIVE_CONSONANTS = [ tɬʼ, cʎ̝̊ʼ, kʟ̝̊ʼ ]
set EJECTIVE_CONSONANTS = { STOP_EJECTIVE_CONSONANTS or FRICATIVE_EJECTIVE_CONSONANTS
or L_FRICATIVE_EJECTIVE_CONSONANTS or AFFRICATE_EJECTIVE_CONSONANTS
or L_AFFRICATE_EJECTIVE_CONSONANTS
}
set TENUIS_CLICK_CONSONANTS = [ ʘ, ǀ, ǃ, ǂ ],
VOICED_CLICK_CONSONANTS = [ ʘ̬, ǀ̬, ǃ̬, ǂ̬ ],
NASAL_CLICK_CONSONANTS = [ ʘ̃, ǀ̃, ǃ̃, ǂ̃ ],
L_CLICK_CONSONANTS = [ ǁ, ǁ̬ ]
set CLICK_CONSONANTS = { TENUIS_CLICK_CONSONANTS or VOICED_CLICK_CONSONANTS
or NASAL_CLICK_CONSONANTS or L_CLICK_CONSONANTS
}
set IMPLOSIVE_CONSONANTS = [ ɓ, ɗ, ᶑ, ʄ, ɠ, ʛ, ɓ̥, ɗ̥, ᶑ̊, ʄ̊, ɠ̊, ʛ̥ ]
set NON_PULMONIC_CONSONANTS = { EJECTIVE_CONSONANTS or CLICK_CONSONANTS or IMPLOSIVE_CONSONANTS }
set CONSONANTS = { PULMONIC_CONSONANTS or NON_PULMONIC_CONSONANTS }
set MODAL_VOWELS = [ i, y, ɨ, ʉ, ɯ, u, ɪ, ʏ, ʊ, e, ø ɘ, ɵ ɤ, o, ø̞ ə, o̞, ɛ, œ ɜ, ɞ ʌ, ɔ, æ, ɐ, a, ɶ, ä, ɑ, ɒ ],
BREATHY_VOWELS = { [ V ] in MODAL_VOWELS yield [ V̤ ] },
VOICELESS_VOWELS = { [ V ] in MODAL_VOWELS yield [ V̥ ] },
CREAKY_VOWELS = { [ V ] in MODAL_VOWELS yield [ V̰ ] }
set SHORT_ORAL_VOWELS = { MODAL_VOWELS or BREATHY_VOWELS or CREAKY_VOWELS or VOICELESS_VOWELS },
LONG_ORAL_VOWELS = { [ V ] in SHORT_ORAL_VOWELS [ Vː ] },
ORAL_VOWELS = { SHORT_ORAL_VOWELS or LONG_ORAL_VOWELS }
set NASAL_VOWELS = { [ V ] in ORAL_VOWELS yield [ Ṽ ] },
SHORT_NASAL_VOWELS = { [ Vː ] in NASAL_VOWELS yield [ V ]ː },
LONG_NASAL_VOWELS = { [ Vː ] in NASAL_VOWELS }
set VOWELS = { ORAL_VOWELS or NASAL_VOWELS }
set PHONES = { VOWELS or CONSONANTS }
; print [ GLOBAL ]
[lateral
+=
L_AFFRICATE_EJECTIVE_CONSONANTS, L_AFFRICATE_PULMONIC_CONSONANTS, L_APPROXIMANT_PULMONIC_CONSONANTS,
L_CLICK_CONSONANTS, L_FRICATIVE_EJECTIVE_CONSONANTS, L_FRICATIVE_PULMONIC_CONSONANTS, L_TAP_PULMONIC_CONSONANTS
-=
{ not { [+ lateral ] in CONSONANTS } }, VOWELS
; alternative
; { not { [+ lateral ] in PHONES } }
]
*proto-lang
|child-lang

View file

@ -1,644 +0,0 @@
; -------- GA ENGLISH PHONETIC INVENTORY
; ---- VOWELS = æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟
; -- NASAL = æ̃ / ẽ / ə̃ / ɑ̃ / ɔ̃ / ɪ̃ / ɛ̃ / ʌ̃ / ʊ̃ / ĩ / ũ
; ɪ̞ / ʊ̞ = lowered
; u̟ = advanced
; -- LABIAL = u̟ / ʊ̞ / ɔ
; -- +HIGH = i / u̟ / ʊ̞ / ɪ̞
; -- -HIGH = ɑ / æ / e / ə / ɛ / ʌ
; -- +LOW = ɑ / æ / ɛ
; -- -LOW = i / u̟ / ʊ̞ / ɪ̞ / e / ə / ʌ
; -- +BACK = ɑ / ɔ / ʌ / ʊ̞ / u̟
; -- -BACK = æ / e / ə / ɪ̞ / ɛ / i
; -- +TENSE = e / i / u̟ / ɑ
; -- -TENSE = æ / ə / ɪ̞ / ɛ / ʌ / ʊ̞ / ɔ
; ---- DIPHTHONGS = eə / eɪ̯ / ju̟ / äɪ̞ / ɔɪ̞ / oʊ̞ / aʊ̞ / ɑɹ / iɹ / ɛɹ / ɔɹ / ʊɹ
; ---- CONSONANTS = p (pʰ) / b (b̥) / t (tʰ)(ɾ)(ʔ) / d (d̥)(ɾ) / tʃ / dʒ (d̥ʒ̊) / k (kʰ) / g (g̊) / f / v (v̥) / θ / ð (ð̥) /
; s / z (z̥) / ʃ / ʒ (ʒ̊) / h (ɦ)(ç) / m (ɱ)(m̩) / n(n̩) / ŋ / l (l̩)/ ɹ (ɹʲ ~ ɹˤ)(ɹ̩) / w (w̥) / j / x / ʔ
; -- PLOSIVES = p / p' / pʰ / t / t' / tʰ ɾ / k / k' / kʰ
; -- AFFRICATES = tʃ / dʒ
; -- FRICATIVES = f / v / θ / ð / s / z / ʃ / ʒ / ç / x
; -- NASAL OBSTRUENTS = m ɱ / n / ŋ
; -- LIQUIDS = l
; -- RHOTIC LIQUIDS = ɹ ɹʲ ɹˤ
; -- SYLLABIC CONSONANTS = m̩ / n̩ / l̩ / ɹ̩
; -- GLIDES = j / w
; -- LARYNGEALS = h ɦ / ʔ [- consonantal sonorant +/- LARYNGEAL FEATURES] only
; -------- distinctive groups
set PLOSIVES = [ p, pʰ, t, tʼ, tʰ, ɾ, kʼ, k, kʰ ]
AFFRICATES = [ tʃʰ, dʒ ]
FRICATIVES = [ f, v, θ, ð, s, z, ʃ, ʒ, ç, x ]
NASALS = [ m, ɱ, n, ŋ ]
LIQUIDS = [ l, ɹ, ɹʲ, ɹˤ ]
SYLLABICS = [ m̩, n̩, l̩, ɹ̩ ]
VOWELS = [ æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟ ]
GLIDES = [ j, w ]
LARYNGEALS = [ h, ɦ, ʔ ]
VOWELS = [ æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟ ]
; ---- implicit
; GLOBAL { all sets }
; ---- set join operations non-mutable!
; { SET_A not SET_B } left anti join
; { SET_A and SET_B } inner join
; { SET_A or SET_B } full outer join
; { not SET_A } = { GLOBAL not SET_A }
; ---- unnecessary sugar
; { not SET_A nor SET_B } = { GLOBAL not { SET_A or SET_B } }
; ---- set character operations - non-mutable!
; { [ Xy ] in SET_A } FILTER: where X is any character and y is a filtering character
; { SET_A yield [ Xy ] } CONCATENATE: performs transformation with (prepended or) appended character
; { SET_A yield [ X concat y ] }
; { SET_A yield [ y concat X ] }
; { SET_A yield y[ X ] } DISSOCIATE: performs transformation removing prepended (or appended) character
; { SET_A yield y dissoc [ X ] }
; { SET_A yield [ X ] dissoc y }
; { [ Xy ] in SET_A yield [ X ]y } combined FILTER and DISSOCIATE
; ---- TENTATIVE!
; ---- set feature operations - non-mutable!
; { [ + feature1 - feature2 ] in SET_A } FILTER: where feature1 and feature2 are filtering features
; { SET_A yield [ X + feature1 ] } TRANSFORMATION: performs transformation with (prepended or) appended character
; { SET_A yield [ X - feature1 ] }
; { SET_A yield [ X - feature1 + feature2 ] }
; { [ X + feature1 - feature2 ] in SET_A yield [ - feature1 + feature2 ] } combined FILTER and TRANSFORMATION
; ---- MAPPING
set PLOSIVES = [ p, t, k ],
FRICATIVES = [ f, s, x ],
; pairs PLOSIVES with FRICATIVES that have matching features = [ pf, ts, kx ]
AFFRICATES = { PLOSIVES yield [ X concat { [ [ X ] - fricative ] in FRICATIVES } ] }
; ---- example with join, character, and feature operations
; set SET_C = { [ PHONE +feature1 ] in { SET_A or SET_B } yield [ PHONE concat y ] }
; -------- main class features
[consonantal
+=
PLOSIVES, AFFRICATES, FRICATIVES, NASALS, LIQUIDS, SYLLABICS
-=
VOWELS, GLIDES, LARYNGEALS
]
[sonorant
+=
VOWELS, GLIDES, LIQUIDS, NASALS, SYLLABICS
-=
PLOSIVES, AFFRICATES, FRICATIVES, LARYNGEALS
]
[approximant
+=
VOWELS, LIQUIDS, GLIDES,
; SYLLABIC LIQUIDS
l̩, ɹ̩
-=
PLOSIVES, AFFRICATES, FRICATIVES, NASALS,
; SYLLABIC NASALS
m̩, n̩
]
; -------- laryngeal features
[voice
+=
VOWELS, GLIDES, LIQUIDS, NASALS, SYLLABICS,
; VOICED FRICATIVES
v, ð, z, ʒ,
; VOICED AFFRICATES
dʒ,
; VOICED LARYNGEALS
ɦ
-=
PLOSIVES,
; VOICELESS AFFRICATES
tʃ,
; VOICELESS FRICATIVES
f, θ, s, ʃ, ç, x,
; VOICELESS LARYNGEALS
h, ʔ
]
[spreadGlottis
+=
; ASPIRATED PLOSIVES
pʰ, tʰ, kʰ,
; ASPIRATED AFFRICATES
; SPREAD LARYNGEALS
h ɦ
-=
VOWELS, FRICATIVES, NASALS, LIQUIDS, SYLLABICS, GLIDES,
; UNASPIRATED PLOSIVES
p, pʼ, t, tʼ, ɾ, k, kʼ,
; UNASPIRATED AFFRICATES
tʃ, dʒ,
; CONSTRICTED LARYNGEALS
ʔ
]
[constrictedGlottis
+=
; LARYNGEALIZED RHOTIC
ɹˤ,
; CONSTRICTED LARYNGEAL
ʔ,
; EJECTIVE PLOSIVES
pʼ, tʼ, kʼ
-=
VOWELS, AFFRICATES, FRICATIVES, NASALS, SYLLABICS, GLIDES,
; UNCONSTRICTED PLOSIVES
{ PLOSIVES not [ p', t', k' ] },
; NON-CONSTRICTED LIQUIDS
l, ɹ ɹʲ,
; SPREAD LARYNGEALS
h ɦ,
]
; -------- manner features
[continuant
+=
; FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ, ç, x,
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; LIQUIDS + RHOTICS
l, ɹ ɹʲ ɹˤ,
; GLIDES
j, w,
; SYLLABIC LIQUIDS
l̩, ɹ̩,
; TAPS
ɾ
-=
; NON-TAP PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ, k, kʼ, kʰ,
; AFFRICATES
tʃ, dʒ,
; NASALS
m ɱ, n, ŋ,
; SYLLABIC NASALS
m̩, n̩
]
[nasal
+=
; NASALS
m ɱ, n, ŋ,
; SYLLABIC NASALS
m̩, n̩
-=
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ, ç, x,
; LIQUIDS + RHOTICS
l, ɹ ɹʲ ɹˤ,
; GLIDES
j, w,
; SYLLABIC LIQUIDS
l̩, ɹ̩,
; PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ, k, kʼ, kʰ,
; AFFRICATES
tʃ, dʒ,
]
[strident
+=
; STRIDENT FRICATIVES
f, v, s, z, ʃ, ʒ,
; STRIDENT AFFRICATES
tʃ, dʒ
-=
; VOWELS
æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ, k, kʼ, kʰ,
; NON-STRIDENT FRICATIVES
θ, ð, ç, x,
; NASAL OBSTRUENTS
m ɱ, n, ŋ,
; RHOTICS + LIQUIDS
l, ɹ ɹʲ ɹˤ,
; SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩,
; GLIDES
j, w
]
[lateral
+=
; LATERAL LIQUIDS
l,
; SYLLABIC LATERALS,
-=
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ, k, kʼ, kʰ
; AFFRICATES
tʃ, dʒ
; FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ, ç, x
; NASAL OBSTRUENTS
m ɱ, n, ŋ
; RHOTIC LIQUIDS
ɹ ɹʲ ɹˤ
; NON-LIQUID SYLLABIC CONSONANTS
m̩, n̩, ɹ̩
; GLIDES
j, w
]
; -------- ---- PLACE features
; -------- labial features
[labial
+=
; ROUNDED VOWELS
u̟, ʊ̞, ɔ, ʊ̃, ũ, ɔ̃
; LABIAL PLOSIVES
p, pʼ, pʰ,
; LABIAL FRICATIVES
f, v,
; LABIAL NASALS
m ɱ,
; LABIAL SYLLABIC CONSONANTS
m̩,
; LABIAL GLIDES
w
-=
; UNROUNDED VOWELS
æ, e, ə, ɑ, ɪ̞, ɛ, ʌ, i, æ̃, ẽ, ə̃, ɑ̃, ɪ̃, ɛ̃, ʌ̃, ĩ,
; NON-LABIAL PLOSIVES
t, tʼ, tʰ ɾ, k, kʼ, kʰ,
; NON-LABIAL AFFRICATES
tʃ, dʒ,
; NON-LABIAL FRICATIVES
θ, ð, s, z, ʃ, ʒ, ç, x,
; NON-LABIAL NASAL OBSTRUENTS
n, ŋ,
; LIQUIDS
l,
; RHOTIC LIQUIDS
ɹ ɹʲ ɹˤ,
; NON-LABIAL SYLLABIC CONSONANTS
n̩, l̩, ɹ̩,
; NON-LABIAL GLIDES
j
]
; -------- coronal features
[coronal
+=
; CORONAL PLOSIVES
t, tʼ, tʰ ɾ,
; CORONAL AFFRICATES
tʃ, dʒ,
; CORONAL FRICATIVES
θ, ð, s, z, ʃ, ʒ,
; CORONAL NASALS
n,
; CORONAL LIQUIDS
l
; CORONAL RHOTIC LIQUIDS
ɹ
; CORONAL SYLLABIC CONSONANTS
n̩, l̩, ɹ̩
-=
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; NON-CORONAL PLOSIVES
p, pʼ, pʰ, k, kʼ, kʰ
; NON-CORONAL FRICATIVES
f, v, ç, x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ, ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩,
; NON-CORONAL GLIDES
j, w
]
[anterior
+=
; ALVEOLAR PLOSIVES
t, tʼ, tʰ ɾ,
; ALVEOLAR AFFRICATES
tʃ, dʒ,
; DENTAL FRICATIVES
θ, ð,
; ALVEOLAR FRICATIVES
s, z,
; ALVEOLAR NASALS
n,
; ALVEOLAR LIQUIDS
l
; ALVEOLAR SYLLABIC CONSONANTS
n̩, l̩,
-=
; POSTALVEOLAR FRICATIVES
ʃ, ʒ,
; POSTALVEOLAR RHOTIC LIQUIDS
ɹ,
; POSTALVEOLAR SYLLABIC CONSONANTS
ɹ̩,
; -- NON-CORONALs
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; NON-CORONAL PLOSIVES
p, pʼ, pʰ, k, kʼ, kʰ
; NON-CORONAL FRICATIVES
f, v, ç, x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ, ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩,
; NON-CORONAL GLIDES
j, w
]
[distributed
+=
; DENTAL FRICATIVES
θ, ð,
; POSTALVEOLAR FRICATIVES
ʃ, ʒ,
; POSTALVEOLAR RHOTIC LIQUIDS
ɹ,
; POSTALVEOLAR SYLLABIC CONSONANTS
ɹ̩,
-=
; apical, retroflex
; ALVEOLAR PLOSIVES
t, tʼ, tʰ ɾ,
; ALVEOLAR FRICATIVES
s, z,
; ALVEOLAR NASALS
n,
; ALVEOLAR LIQUIDS
l
; ALVEOLAR SYLLABIC CONSONANTS
n̩, l̩,
; -- NON-CORONALS
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; NON-CORONAL PLOSIVES
p, pʼ, pʰ, k, kʼ, kʰ
; NON-CORONAL FRICATIVES
f, v, ç, x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ, ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩,
; NON-CORONAL GLIDES
j, w
]
; -------- dorsal features
[dorsal
+=
; VOWELS
æ, e, ə, ɑ, ɔ, ɪ̞, ɛ, ʌ, ʊ̞, i, u̟, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃, ĩ, ũ
; DORSAL PLOSIVES
k, kʼ, kʰ,
; DORSAL FRICATIVES
ç, x,
; DORSAL NASAL OBSTRUENTS
ŋ,
; DORSAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; DORSAL GLIDES
j
-=
; NON-DORSAL PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ,
; NON-DORSAL AFFRICATES
tʃ, dʒ,
; NON-DORSAL FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ,
; NON-DORSAL NASALS
m ɱ, n,
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩
; NON-DORSAL GLIDES
w
]
[high
+=
; HIGH VOWELS
i, u̟, ʊ̞, ɪ̞, ĩ, ũ, ʊ̃, ɪ̃
; HIGH DORSAL PLOSIVES
k, kʼ, kʰ,
; HIGH DORSAL FRICATIVES
ç, x,
; HIGH DORSAL NASAL OBSTRUENTS
ŋ,
; HIGH RHOTIC LIQUIDS
ɹʲ
; HIGH DORSAL GLIDES
j, w
-= χ, e, o, a
; NON-HIGH VOWELS
ɑ, æ, e, ə, ɛ, ʌ, æ̃, ẽ, ə̃, ɑ̃, ɔ̃, ɛ̃, ʌ̃,
; NON-HIGH RHOTIC LIQUIDS
ɹˤ
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ,
; NON-DORSAL AFFRICATES
tʃ, dʒ,
; NON-DORSAL FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ,
; NON-DORSAL NASALS
m ɱ, n,
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩
; NON-DORSAL GLIDES
w
]
[low
+=
; LOW VOWELS
ɑ, æ, ɛ, æ̃, ɑ̃, ɛ̃,
; LOW DORSAL RHOTIC LIQUIDS
ɹˤ
-= a, ɛ, ɔ
; NON-LOW VOWELS
i, u̟, ʊ̞, ɪ̞, e, ə, ʌ, ẽ, ə̃, ɔ̃, ɪ̃, ʌ̃, ʊ̃, ĩ, ũ
; NON-LOW DORSAL PLOSIVES
k, kʼ, kʰ,
; NON-LOW DORSAL FRICATIVES
ç, x,
; NON-LOW DORSAL NASAL OBSTRUENTS
ŋ,
; NON-LOW DORSAL RHOTIC LIQUIDS
ɹʲ
; DORSAL GLIDES
j
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ,
; NON-DORSAL AFFRICATES
tʃ, dʒ,
; NON-DORSAL FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ,
; NON-DORSAL NASALS
m ɱ, n,
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩
; NON-DORSAL GLIDES
w
]
[back
+=
; k, kʼ, ɣ, χ, u, ə, o, ʌ, ɑ
; BACK VOWELS
ɑ, ɔ, ʌ, ʊ̞, u̟, ɑ̃, ɔ̃, ʌ̃, ʊ̃, ũ,
; BACK DORSAL PLOSIVES
k, kʼ, kʰ,
; BACK DORSAL FRICATIVES
x,
; BACK DORSAL NASAL OBSTRUENTS
ŋ,
; BACK DORSAL RHOTIC LIQUIDS
ɹˤ
-= ç, k̟, i, y, ø, ɛ
; NON-BACK DORSAL FRICATIVES
ç,
; NON-BACK DORSAL RHOTIC LIQUIDS
ɹʲ
; NON-BACK DORSAL GLIDES
j
; NON-BACK VOWELS
æ, e, ə, ɪ̞, ɛ, i, æ̃, ẽ, ə̃, ɪ̃, ɛ̃, ĩ
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ,
; NON-DORSAL AFFRICATES
tʃ, dʒ,
; NON-DORSAL FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ,
; NON-DORSAL NASALS
m ɱ, n,
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩
; NON-DORSAL GLIDES
w
]
[tense ; compare to ATR or RTR
+=
; TENSE VOWELS
e, i, u̟, ɑ, ĩ, ũ, ẽ, ɑ̃,
-=
; NON-TENSE VOWELS
æ, ə, ɪ̞, ɛ, ʌ, ʊ̞, ɔ, æ̃, ə̃, ɔ̃, ɪ̃, ɛ̃, ʌ̃, ʊ̃,
; DORSAL PLOSIVES
k, kʼ, kʰ,
; DORSAL FRICATIVES
ç, x,
; DORSAL NASAL OBSTRUENTS
ŋ,
; DORSAL RHOTIC LIQUIDS
ɹʲ ɹˤ,
; DORSAL GLIDES
j
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p, pʼ, pʰ, t, tʼ, tʰ ɾ,
; NON-DORSAL AFFRICATES
tʃ, dʒ,
; NON-DORSAL FRICATIVES
f, v, θ, ð, s, z, ʃ, ʒ,
; NON-DORSAL NASALS
m ɱ, n,
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩, n̩, l̩, ɹ̩
; NON-DORSAL GLIDES
w
]
*PROTO
|Gif Lang
*PROTO
|Jif Lang
; -- Devoicing, all our z's become s's
[ + voice consonantal - nasal]>[- voice]/._.
; -- loss of schwa, the is th'
ə>0/._.
; -- Ejectivization, all our pits become pit's
[+ spreadGlottis - continuant]>[+ constrictedGlottis - spreadGlottis]/._[+ constrictedGlottis]
[+ spreadGlottis - continuant]>[+ constrictedGlottis - spreadGlottis]/[+ constrictedGlottis]_.
[+ constrictedGlottis]>0/[+ constrictedGlottis - continuant]_.
[+ constrictedGlottis]>0/._[+ constrictedGlottis - continuant]
; -- r color spreading, all our reports become rihpahts
[- consonantal tense]>[+ tense]/ɹ_.
[- consonantal tense]>[+ tense]/._ɹ
[- consonantal high]>[+ high]/ɹʲ_.
[- consonantal high]>[+ high]/._ɹʲ
[- consonantal back]>[+ back]/ɹˤ_.
[- consonantal back]>[+ back]/._ɹˤ
ɹ>0/._.
ɹʲ>0/._.
ɹˤ>0/._.
; -- Deaspiration, tiff is diff and diff is tiff
[+ spreadGlottis - continuant]>[- spreadGlottis]/._.
; "JavaScript"
; "gif or jif? I say zhaif"
; "This request returns an empty object"
; "I love going to waffle js!"
; "A donut a day makes living with the threat of pandemic easier"

BIN
public/logo192.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 8.4 KiB

BIN
public/logo512.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 22 KiB

View file

@ -1,11 +1,21 @@
{ {
"short_name": "FCA", "short_name": "React App",
"name": "Feature Change Applier", "name": "Create React App Sample",
"icons": [ "icons": [
{ {
"src": "favicon.ico", "src": "favicon.ico",
"sizes": "64x64 32x32 24x24 16x16", "sizes": "64x64 32x32 24x24 16x16",
"type": "image/x-icon" "type": "image/x-icon"
},
{
"src": "logo192.png",
"type": "image/png",
"sizes": "192x192"
},
{
"src": "logo512.png",
"type": "image/png",
"sizes": "512x512"
} }
], ],
"start_url": ".", "start_url": ".",

View file

@ -1,7 +0,0 @@
$colors: (
"main--bg": #281734,
"main": #d5bfbf,
"text-input": #e8e22e,
"text-input--bg": #1d191a,
"error": #ff0000
);

View file

@ -1,366 +0,0 @@
/* http://meyerweb.com/eric/tools/css/reset/
v2.0-modified | 20110126
License: none (public domain)
*/
html, body, div, span, applet, object, iframe,
h1, h2, h3, h4, h5, h6, p, blockquote, pre,
a, abbr, acronym, address, big, cite, code,
del, dfn, em, img, ins, kbd, q, s, samp,
small, strike, strong, sub, sup, tt, var,
b, u, i, center,
dl, dt, dd, ol, ul, li,
fieldset, form, label, legend,
table, caption, tbody, tfoot, thead, tr, th, td,
article, aside, canvas, details, embed,
figure, figcaption, footer, header, hgroup,
menu, nav, output, ruby, section, summary,
time, mark, audio, video {
margin: 0;
padding: 0;
border: 0;
font-size: 100%;
font: inherit;
vertical-align: baseline;
}
/* make sure to set some focus styles for accessibility */
:focus {
outline: 0;
}
/* HTML5 display-role reset for older browsers */
article, aside, details, figcaption, figure,
footer, header, hgroup, menu, nav, section {
display: block;
}
body {
line-height: 1;
}
ol, ul {
list-style: none;
}
blockquote, q {
quotes: none;
}
blockquote:before, blockquote:after,
q:before, q:after {
content: '';
content: none;
}
table {
border-collapse: collapse;
border-spacing: 0;
}
input[type=search]::-webkit-search-cancel-button,
input[type=search]::-webkit-search-decoration,
input[type=search]::-webkit-search-results-button,
input[type=search]::-webkit-search-results-decoration {
-webkit-appearance: none;
-moz-appearance: none;
}
input[type=search] {
-webkit-appearance: none;
-moz-appearance: none;
-webkit-box-sizing: content-box;
-moz-box-sizing: content-box;
box-sizing: content-box;
}
textarea {
overflow: auto;
vertical-align: top;
resize: vertical;
}
/**
* Correct `inline-block` display not defined in IE 6/7/8/9 and Firefox 3.
*/
audio,
canvas,
video {
display: inline-block;
*display: inline;
*zoom: 1;
max-width: 100%;
}
/**
* Prevent modern browsers from displaying `audio` without controls.
* Remove excess height in iOS 5 devices.
*/
audio:not([controls]) {
display: none;
height: 0;
}
/**
* Address styling not present in IE 7/8/9, Firefox 3, and Safari 4.
* Known issue: no IE 6 support.
*/
[hidden] {
display: none;
}
/**
* 1. Correct text resizing oddly in IE 6/7 when body `font-size` is set using
* `em` units.
* 2. Prevent iOS text size adjust after orientation change, without disabling
* user zoom.
*/
html {
font-size: 100%; /* 1 */
-webkit-text-size-adjust: 100%; /* 2 */
-ms-text-size-adjust: 100%; /* 2 */
}
/**
* Address `outline` inconsistency between Chrome and other browsers.
*/
a:focus {
outline: thin dotted;
}
/**
* Improve readability when focused and also mouse hovered in all browsers.
*/
a:active,
a:hover {
outline: 0;
}
/**
* 1. Remove border when inside `a` element in IE 6/7/8/9 and Firefox 3.
* 2. Improve image quality when scaled in IE 7.
*/
img {
border: 0; /* 1 */
-ms-interpolation-mode: bicubic; /* 2 */
}
/**
* Address margin not present in IE 6/7/8/9, Safari 5, and Opera 11.
*/
figure {
margin: 0;
}
/**
* Correct margin displayed oddly in IE 6/7.
*/
form {
margin: 0;
}
/**
* Define consistent border, margin, and padding.
*/
fieldset {
border: 1px solid #c0c0c0;
margin: 0 2px;
padding: 0.35em 0.625em 0.75em;
}
/**
* 1. Correct color not being inherited in IE 6/7/8/9.
* 2. Correct text not wrapping in Firefox 3.
* 3. Correct alignment displayed oddly in IE 6/7.
*/
legend {
border: 0; /* 1 */
padding: 0;
white-space: normal; /* 2 */
*margin-left: -7px; /* 3 */
}
/**
* 1. Correct font size not being inherited in all browsers.
* 2. Address margins set differently in IE 6/7, Firefox 3+, Safari 5,
* and Chrome.
* 3. Improve appearance and consistency in all browsers.
*/
button,
input,
select,
textarea {
font-size: 100%; /* 1 */
margin: 0; /* 2 */
vertical-align: baseline; /* 3 */
*vertical-align: middle; /* 3 */
}
/**
* Address Firefox 3+ setting `line-height` on `input` using `!important` in
* the UA stylesheet.
*/
button,
input {
line-height: normal;
}
/**
* Address inconsistent `text-transform` inheritance for `button` and `select`.
* All other form control elements do not inherit `text-transform` values.
* Correct `button` style inheritance in Chrome, Safari 5+, and IE 6+.
* Correct `select` style inheritance in Firefox 4+ and Opera.
*/
button,
select {
text-transform: none;
}
/**
* 1. Avoid the WebKit bug in Android 4.0.* where (2) destroys native `audio`
* and `video` controls.
* 2. Correct inability to style clickable `input` types in iOS.
* 3. Improve usability and consistency of cursor style between image-type
* `input` and others.
* 4. Remove inner spacing in IE 7 without affecting normal text inputs.
* Known issue: inner spacing remains in IE 6.
*/
button,
html input[type="button"], /* 1 */
input[type="reset"],
input[type="submit"] {
-webkit-appearance: button; /* 2 */
cursor: pointer; /* 3 */
*overflow: visible; /* 4 */
}
/**
* Re-set default cursor for disabled elements.
*/
button[disabled],
html input[disabled] {
cursor: default;
}
/**
* 1. Address box sizing set to content-box in IE 8/9.
* 2. Remove excess padding in IE 8/9.
* 3. Remove excess padding in IE 7.
* Known issue: excess padding remains in IE 6.
*/
input[type="checkbox"],
input[type="radio"] {
box-sizing: border-box; /* 1 */
padding: 0; /* 2 */
*height: 13px; /* 3 */
*width: 13px; /* 3 */
}
/**
* 1. Address `appearance` set to `searchfield` in Safari 5 and Chrome.
* 2. Address `box-sizing` set to `border-box` in Safari 5 and Chrome
* (include `-moz` to future-proof).
*/
input[type="search"] {
-webkit-appearance: textfield; /* 1 */
-moz-box-sizing: content-box;
-webkit-box-sizing: content-box; /* 2 */
box-sizing: content-box;
}
/**
* Remove inner padding and search cancel button in Safari 5 and Chrome
* on OS X.
*/
input[type="search"]::-webkit-search-cancel-button,
input[type="search"]::-webkit-search-decoration {
-webkit-appearance: none;
}
/**
* Remove inner padding and border in Firefox 3+.
*/
button::-moz-focus-inner,
input::-moz-focus-inner {
border: 0;
padding: 0;
}
/**
* 1. Remove default vertical scrollbar in IE 6/7/8/9.
* 2. Improve readability and alignment in all browsers.
*/
textarea {
overflow: auto; /* 1 */
vertical-align: top; /* 2 */
}
/**
* Remove most spacing between table cells.
*/
table {
border-collapse: collapse;
border-spacing: 0;
}
html,
button,
input,
select,
textarea {
color: #222;
}
::-moz-selection {
background: #b3d4fc;
text-shadow: none;
}
::selection {
background: #b3d4fc;
text-shadow: none;
}
img {
vertical-align: middle;
}
fieldset {
border: 0;
margin: 0;
padding: 0;
}
textarea {
resize: vertical;
}
.chromeframe {
margin: 0.2em 0;
background: #ccc;
color: #000;
padding: 0.2em 0;
}

View file

@ -5,7 +5,7 @@ import PhonoChangeApplier from './PhonoChangeApplier';
function App() { function App() {
return ( return (
<div className="App" data-testid="App"> <div className="App" data-testid="App">
<h1 data-testid="App-name">Feature Change Applier</h1> <h1>Phono Change Applier</h1>
<PhonoChangeApplier /> <PhonoChangeApplier />
</div> </div>
); );

View file

@ -1,6 +1,5 @@
import React from 'react'; import React from 'react';
import ReactDOM from 'react-dom'; import ReactDOM from 'react-dom';
import { HashRouter as Router } from 'react-router-dom';
import App from './App'; import App from './App';
import renderer from 'react-test-renderer'; import renderer from 'react-test-renderer';
import { exportAllDeclaration } from '@babel/types'; import { exportAllDeclaration } from '@babel/types';
@ -9,13 +8,13 @@ import extendExpect from '@testing-library/jest-dom/extend-expect'
it('renders App without crashing', () => { it('renders App without crashing', () => {
const div = document.createElement('div'); const div = document.createElement('div');
ReactDOM.render(<Router><App /></Router>, div); ReactDOM.render(<App />, div);
ReactDOM.unmountComponentAtNode(div); ReactDOM.unmountComponentAtNode(div);
}); });
describe('App', () => { describe('App', () => {
it('renders the correct title', () => { it('renders the correct title', () => {
const { getByTestId } = render(<Router><App /></Router>); const { getByTestId } = render(<App />);
expect(getByTestId('App-name')).toHaveTextContent('Feature Change Applier'); expect(getByTestId('App')).toHaveTextContent('Phono Change Applier');
}) })
}) })

View file

@ -1,51 +1,133 @@
import React, { useState, useReducer } from 'react'; import React, { useState, useReducer } from 'react';
import { Link, Route } from 'react-router-dom';
import './PhonoChangeApplier.scss'; import './PhonoChangeApplier.scss';
// import ls from 'local-storage';
import ProtoLang from './components/ProtoLang'; import ProtoLang from './components/ProtoLang';
import Features from './components/Features'; import Features from './components/Features';
import Epochs from './components/Epochs'; import Epochs from './components/Epochs';
import Options from './components/Options'; import Options from './components/Options';
import Output from './components/Output'; import Output from './components/Output';
import Latl from './components/Latl'; import {stateReducer} from './reducers/stateReducer';
import LatlOutput from './components/LatlOutput'; import {initState} from './reducers/stateReducer.init';
import { stateReducer } from './reducers/reducer';
import { clearState, waffleState } from './reducers/reducer.init';
const PhonoChangeApplier = () => { const PhonoChangeApplier = () => {
const [ state, dispatch ] = useReducer( const [ state, dispatch ] = useReducer(
stateReducer, stateReducer,
{}, {},
waffleState initState
) )
const { lexicon, phones, phonemes, epochs, options, features, results, errors, latl, parseResults } = state; // ! DONE
const [ lexicon, setLexicon ] = useState(['mun', 'tʰu', 'tɯm', 'utʰ']);
// ! UNDONE
const [ phonemes, setPhonemes ] = useState(
{
n: [ 'occlusive', 'sonorant', 'obstruent', 'nasal', 'alveolar' ],
m: [ 'occlusive', 'sonorant', 'obstruent', 'nasal', 'bilabial' ],
u: [ 'continuant', 'sonorant', 'syllabic', 'high', 'back', 'rounded' ],
ɯ: [ 'continuant', 'sonorant', 'syllabic', 'high', 'back', 'unrounded' ],
t: [ 'occlusive', 'plosive', 'obstruent', 'alveolar' ],
: [ 'occlusive', 'plosive', 'obstruent', 'alveolar', 'aspirated' ],
}
);
const [ epochs, setEpochs ] = useState([{name: 'epoch 1', changes:['[+ rounded]>[- rounded + unrounded]/_#']}]);
const [ options, setOptions ] = useState({output: 'default', save: false})
const [ results, setResults ] = useState([])
const [ errors, setErrors ] = useState({})
const [ features, setFeatures ] = useState(
['occlusive', 'sonorant', 'obstruent', 'nasal', 'alveolar','bilabial',
'continuant','syllabic','high','back','rounded','unrounded', 'plosive','aspirated'])
const runChanges = e => {
e.preventDefault();
let ruleError = epochs.reduce((errorObject, epoch) => {
epoch.changes.map((change, index) => {
if (!change.match(/>.*\/.*_/)) {
errorObject[epoch.name]
? errorObject[epoch.name].push(index)
: errorObject[epoch.name] = [index]
errorObject[epoch.name].ruleSyntaxError = true;
}
// TODO validate phoneme syntax
let decomposedChange = change.split('>');
decomposedChange = [decomposedChange[0], ...decomposedChange[1].split('/')]
decomposedChange = [decomposedChange[0], decomposedChange[1], ...decomposedChange[2].split('_')];
})
return errorObject;
}, {})
if (Object.entries(ruleError).length) return setErrors(ruleError)
setErrors({});
// decompose Lexical Items
// moving window on phonemes of each lexical item
let lexicalFeatureBundles = []
lexicon.forEach(lexeme => {
let lexemeBundle = [];
let startingIndex = 0;
let lastIndex = lexeme.length - 1;
[...lexeme].forEach((_, index) => {
if (phonemes[lexeme.slice(startingIndex, index + 1)] && index !== lastIndex) return;
if (phonemes[lexeme.slice(startingIndex, index + 1)]) return lexemeBundle.push(phonemes[lexeme.slice(startingIndex)])
if (index !== 0 && index !== lastIndex) lexemeBundle.push(phonemes[lexeme.slice(startingIndex, index)])
if (index === lastIndex) {
lexemeBundle.push(phonemes[lexeme.slice(startingIndex, index)])
lexemeBundle.push(phonemes[lexeme.slice(index)])
}
startingIndex = index;
})
lexemeBundle.unshift(['#'])
lexemeBundle.push(['#'])
lexicalFeatureBundles.push(lexemeBundle);
})
console.log(lexicalFeatureBundles)
// decompose rules
let allEpochs = epochs.map(epoch => {
let ruleBundle = epoch.changes.map(rule => {
return {
input: rule.split('>')[0].replace(/\[|\]|\+/g, '').trim(),
result: rule.split('>')[1].split('/')[0],
preInput: rule.split('/')[1].split('_')[0].replace(/\[|\]|\+/g, '').trim(),
postInput: rule.split('/')[1].split('_')[1].replace(/\[|\]|\+/g, '').trim(),
}
})
return {epoch: epoch.name, rules: ruleBundle}
})
console.log(allEpochs)
// apply sound changes
allEpochs.reduce((diachronicLexicon, epoch) => {
let startingLexicon = diachronicLexicon.length
? diachronicLexicon[diachronicLexicon.length - 1]
: lexicalFeatureBundles;
let currentRules = epoch.rules;
let resultingLexicon = startingLexicon.forEach(lexeme => {
currentRules.forEach(rule => {
let ruleEnvironment = [[rule.preInput], [rule.input], [rule.postInput]];
console.log(ruleEnvironment)
})
})
diachronicLexicon.push(resultingLexicon)
},[])
// handle output
}
return ( return (
<> <div className="PhonoChangeApplier" data-testid="PhonoChangeApplier">
<ProtoLang lexicon={state.lexicon} dispatch={dispatch}/>
<Route exact path="/latl"> <Features phones={state.phones} features={state.features} dispatch={dispatch}/>
<Link to="/">Back to GUI</Link> <Epochs epochs={state.epochs} dispatch={dispatch} />
<div className="PhonoChangeApplier PhonoChangeApplier--latl"> <Options options={state.options} dispatch={dispatch}/>
<Latl latl={latl} dispatch={dispatch}/> <Output results={results} setResults={setResults}/>
<LatlOutput results={results} options={options} parseResults={parseResults} errors={errors} dispatch={dispatch}/>
</div> </div>
</Route>
<Route exact path="/">
<Link to="/latl">LATL</Link>
<div className="PhonoChangeApplier PhonoChangeApplier--gui" data-testid="PhonoChangeApplier">
<ProtoLang lexicon={lexicon} dispatch={dispatch}/>
<Features phones={phones} features={features} dispatch={dispatch}/>
<Epochs epochs={epochs} errors={errors} dispatch={dispatch} />
<Options options={options} dispatch={dispatch}/>
<Output results={results} options={options} dispatch={dispatch}/>
</div>
</Route>
</>
); );
} }

View file

@ -1,67 +0,0 @@
@import '../public/stylesheets/variables';
div.App {
max-height: 100vh;
max-width: 100vw;
line-height: 1.25em;
padding: 1em;
a {
color: map-get($colors, 'text-input')
}
h1 {
font-size: 2em;
padding: 1em 0;
}
h3 {
font-size: 1.25em;
padding: 0.5em 0;
}
h5 {
font-size: 1.1em;
padding: 0.1em 0;
font-weight: 800;
}
div.PhonoChangeApplier--gui {
display: grid;
width: 100%;
place-items: center center;
grid-template-columns: repeat(auto-fit, minmax(25em, 1fr));
grid-template-rows: repeat(auto-fill, minmax(300px, 1fr));
div {
max-width: 100%;
max-height: 50vh;
margin: 1em;
overflow-y: scroll;
}
}
div.PhonoChangeApplier--latl {
display: flex;
flex-flow: row wrap;
}
button.form, input[type="submit"].form, input[type="button"].form {
height: 2em;
border-radius: 0.25em;
border-color: transparent;
margin: 0.2em auto;
width: 10em;
}
button.form--add, input[type="submit"].form--add, input[type="button"].form--add{
background-color: greenyellow;
color: black;
}
button.form--remove, input[type="submit"].form--remove, input[type="button"].form--remove {
background-color: red;
color: white;
}
}

View file

@ -1,6 +1,5 @@
import React from 'react'; import React from 'react';
import ReactDOM from 'react-dom'; import ReactDOM from 'react-dom';
import { HashRouter as Router } from 'react-router-dom';
import App from './App'; import App from './App';
import PhonoChangeApplier from './PhonoChangeApplier'; import PhonoChangeApplier from './PhonoChangeApplier';
import renderer from 'react-test-renderer'; import renderer from 'react-test-renderer';
@ -10,13 +9,13 @@ import extendExpect from '@testing-library/jest-dom/extend-expect'
it('renders PhonoChangeApplier without crashing', () => { it('renders PhonoChangeApplier without crashing', () => {
const div = document.createElement('div'); const div = document.createElement('div');
ReactDOM.render(<Router><PhonoChangeApplier /></Router>, div); ReactDOM.render(<PhonoChangeApplier />, div);
ReactDOM.unmountComponentAtNode(div); ReactDOM.unmountComponentAtNode(div);
}); });
describe('App', () => { describe('App', () => {
it('renders Proto Language Lexicon', () => { it('renders Proto Language Lexicon', () => {
const { getByTestId } = render(<Router><PhonoChangeApplier /></Router>); const { getByTestId } = render(<PhonoChangeApplier />);
expect(getByTestId('PhonoChangeApplier')).toHaveTextContent('Proto Language Lexicon'); expect(getByTestId('PhonoChangeApplier')).toHaveTextContent('Proto Language Lexicon');
}) })
}) })

View file

@ -2,77 +2,52 @@ import React from 'react';
import './Epochs.scss'; import './Epochs.scss';
import SoundChangeSuite from './SoundChangeSuite'; import SoundChangeSuite from './SoundChangeSuite';
import { render } from 'react-dom';
const Epochs = ({epochs, errors, dispatch}) => { const Epochs = props => {
const addEpoch = e => {
const addEpoch = (e, props) => {
e.preventDefault() e.preventDefault()
let index = epochs.length + 1; let index = props.epochs.length + 1;
dispatch({ props.dispatch({
type: 'ADD_EPOCH', type: 'ADD_EPOCH',
value: {name: `epoch ${index}`} value: {name: `Epoch ${index}`}
}) })
} }
const removeEpoch = (e, epochName) => { const removeEpoch = (e, epochName) => {
e.preventDefault() e.preventDefault()
dispatch({ props.dispatch({
type: 'REMOVE_EPOCH', type: 'REMOVE_EPOCH',
value: {name: epochName} value: {name: epochName}
}); });
} }
const updateEpoch = (epoch, epochIndex) => { const updateEpoch = (epoch, epochIndex) => {
const dispatchValue = { let updatedEpochs = [...props.epochs]
name: epoch.name, updatedEpochs[epochIndex] = epoch
index: epochIndex, props.dispatch({
changes: epoch.changes,
parent: epoch.parent
}
dispatch({
type: "SET_EPOCH", type: "SET_EPOCH",
value: dispatchValue value: epoch
}) })
} }
const renderAddEpochButton = index => {
if (epochs && index === epochs.length - 1 ) return (
<form onSubmit={e=>addEpoch(e)}>
<input className="form form--add" type="submit" name="add-epoch" value="Add Epoch" ></input>
</form>
)
return <></>
}
const renderEpochs = () => {
if (epochs && epochs.length) {
return epochs.map((epoch, index) => {
const epochError = errors.epoch ? errors.error : null
return ( return (
<div <div className="Epochs" data-testid="Epochs">
className="SoundChangeSuite" <h3>Sound Change Epochs</h3>
data-testid={`${epoch.name}_SoundChangeSuite`} {props.epochs
key={`epoch-${index}`} ? props.epochs.map((epoch, idx) => {
> return <SoundChangeSuite
<SoundChangeSuite key={`epochname-${idx}`} epochIndex={idx} epoch={epoch}
epochIndex={index} epoch={epoch}
updateEpoch={updateEpoch} removeEpoch={removeEpoch} updateEpoch={updateEpoch} removeEpoch={removeEpoch}
epochs={epochs} // error={props.errors[epoch.name]}
error={epochError} />})
/> : <></>}
{renderAddEpochButton(index)} <form onSubmit={e=>addEpoch(e, props)}>
<input type="submit" name="add-epoch" value="Add Epoch" ></input>
</form>
</div> </div>
)});
}
return renderAddEpochButton(-1)
}
return (
<>
{ renderEpochs() }
</>
); );
} }

View file

@ -13,6 +13,10 @@ it('renders Epochs without crashing', () => {
}); });
describe('Epochs', () => { describe('Epochs', () => {
it('renders the correct subtitle', () => {
const { getByTestId } = render(<Epochs />);
expect(getByTestId('Epochs')).toHaveTextContent('Sound Change Epochs');
});
it('renders a suite of soundchanges', () => { it('renders a suite of soundchanges', () => {
const { getByTestId } = render(<Epochs />); const { getByTestId } = render(<Epochs />);

View file

@ -2,64 +2,27 @@
import React, {useState} from 'react'; import React, {useState} from 'react';
import './Features.scss'; import './Features.scss';
import type { featureAction } from '../reducers/reducer.features'; import type { featureAction } from '../reducers/stateReducer.features';
const parsePhonesFromFeatureObject = featureObject => {
const Features = ({ phones, features, dispatch }) => {
const [feature, setFeature] = useState('aspirated')
const [ newPositivePhones, setNewPositivePhones ] = useState('tʰ / pʰ / kʰ');
const [ newNegativePhones, setNewNegativePhones ] = useState('t / p / k');
const newFeaturesSubmit = e => {
e.preventDefault();
setFeature('');
setNewPositivePhones('');
setNewNegativePhones('');
}
const handleDeleteClick = (e, feature) => {
e.preventDefault();
const deleteFeatureAction = {
type: "DELETE_FEATURE",
value: feature
}
return dispatch(deleteFeatureAction);
}
const parsePhonesFromFeatureObject = featureObject => {
const getProperty = property => object => object[property]
const getFeatureMap = (featureObject) => { const getFeatureMap = (featureObject) => {
return Object.keys(featureObject).map(feature => { return Object.keys(featureObject).map(feature => {
const plusPhones = featureObject[feature].positive.map(getProperty('grapheme')).join(' / '); const plusPhones = featureObject[feature].positive.map(phone => phone.grapheme).join('|');
const minusPhones = featureObject[feature].negative.map(getProperty('grapheme')).join(' / '); const minusPhones = featureObject[feature].negative.map(phone => phone.grapheme).join('|');
return {[feature]: {plus: plusPhones, minus: minusPhones}} return {[feature]: {plus: plusPhones, minus: minusPhones}}
}) })
} }
const getFeatureMapJSX = (featureMap) => { const getFeatureMapJSX = (featureMap) => {
return featureMap.map((feature, index) => { return featureMap.map((feature, index) => {
const [featureName] = Object.keys(feature); const featureName = Object.keys(feature)
const { plus, minus } = feature[featureName]; const plusPhones = feature[featureName].plus;
const minusPhones = feature[featureName].minus;
return ( return (
<li key={`feature__${featureName}`}> <li key={`feature__${featureName}`}>
<span className="feature--names-and-phones"> <span className="plus-phones">{`[+ ${featureName}] = ${plusPhones}`}</span>
<span className="feature--feature-name"> <span className="minus-phones">{`[- ${featureName}] = ${minusPhones}`}</span>
{`[+ ${featureName} ]`}
</span>
<span className="feature--feature-phones">
{plus}
</span>
</span>
<span className="feature--names-and-phones">
<span className="feature--feature-name">
{`[- ${featureName} ]`}
</span>
<span className="feature--feature-phones">
{minus}
</span>
</span>
<button className="delete-feature" onClick={e => handleDeleteClick(e, featureName)}>X</button>
</li> </li>
) )
}) })
@ -68,28 +31,57 @@ const Features = ({ phones, features, dispatch }) => {
const featureMap = getFeatureMap(featureObject); const featureMap = getFeatureMap(featureObject);
const featureMapJSX = getFeatureMapJSX(featureMap); const featureMapJSX = getFeatureMapJSX(featureMap);
return featureMapJSX; return featureMapJSX;
} }
const parseNewPhones = somePhones => { const buildReducerAction = (e, newPositivePhones, newNegativePhones, feature): featureAction => {
if (somePhones === '') return [''];
return somePhones.split('/').map(phone => phone.trim());
}
const handleClickDispatch = e => dispatchFunction => actionBuilder => actionParameters => {
e.preventDefault(); e.preventDefault();
return dispatchFunction(actionBuilder(actionParameters)); const positivePhones = []
} newPositivePhones !== ''
? newPositivePhones.split('/').forEach(phone => positivePhones.push(phone.trim()))
: positivePhones.push('')
const buildAddFeatureAction = ([newPositivePhones, newNegativePhones, feature]): featureAction => ( const negativePhones = []
{ newNegativePhones !== ''
? newNegativePhones.split('/').forEach(phone => negativePhones.push(phone.trim()))
: negativePhones.push('')
return {
type: "ADD_FEATURE", type: "ADD_FEATURE",
value: { value: {
positivePhones: parseNewPhones(newPositivePhones), positivePhones,
negativePhones: parseNewPhones(newNegativePhones), negativePhones,
feature feature
} }
} }
) }
const getPhonemesFromFeatureSubmission = (props, newPhonemes, feature) => {
let newPhonemeObject = newPhonemes.split('/').reduce((phonemeObject, newPhoneme) => {
newPhoneme = newPhoneme.trim();
phonemeObject = phonemeObject[newPhoneme]
? {...phonemeObject, [newPhoneme]: [...phonemeObject[newPhoneme], feature]}
: {...phonemeObject, [newPhoneme]: [feature]}
return phonemeObject;
}, {...props.phonemes})
return newPhonemeObject;
}
const Features = (props) => {
const [feature, setFeature] = useState('aspirated')
const [ newPositivePhones, setNewPositivePhones ] = useState('tʰ / pʰ / kʰ');
const [ newNegativePhones, setNewNegativePhones ] = useState('t / p / k');
const newFeaturesSubmit = e => {
e.preventDefault();
// let newPhonemeObject = getPhonemesFromFeatureSubmission(props, newPhonemes, feature);
// props.setPhonemes(newPhonemeObject);
// if (!props.features || !props.features.includes(feature)) props.setFeatures([...props.features, feature])
setFeature('');
setNewPositivePhones('');
setNewNegativePhones('');
}
return ( return (
<div className="Features" data-testid="Features"> <div className="Features" data-testid="Features">
@ -97,7 +89,7 @@ const Features = ({ phones, features, dispatch }) => {
<h3>Phonetic Features</h3> <h3>Phonetic Features</h3>
<ul className="Features__list" data-testid="Features-list"> <ul className="Features__list" data-testid="Features-list">
{phones ? <>{parsePhonesFromFeatureObject(features)}</> : <></>} {props.phones ? <>{parsePhonesFromFeatureObject(props.features)}</> : <></>}
</ul> </ul>
<form className="Features__form" data-testid="Features-form"> <form className="Features__form" data-testid="Features-form">
@ -107,27 +99,24 @@ const Features = ({ phones, features, dispatch }) => {
></input> ></input>
{/* ! Positive Phones */} {/* ! Positive Phones */}
<label htmlFor="positive-phones">+ <label htmlFor="positive-phones">+</label>
<input <input
id="positive-phones" id="positive-phones"
type="text" name="phonemes" type="text" name="phonemes"
value={newPositivePhones} onChange={e=> setNewPositivePhones(e.target.value)} value={newPositivePhones} onChange={e=> setNewPositivePhones(e.target.value)}
></input> ></input>
</label>
{/* ! Negative Phones */} {/* ! Negative Phones */}
<label htmlFor="negative-phones">- <label htmlFor="negative-phones">-</label>
<input <input
id="negative-phones" id="negative-phones"
type="text" name="phonemes" type="text" name="phonemes"
value={newNegativePhones} onChange={e=> setNewNegativePhones(e.target.value)} value={newNegativePhones} onChange={e=> setNewNegativePhones(e.target.value)}
></input> ></input>
</label>
<input <input
className="form form--add"
type="submit" type="submit"
onClick={e => handleClickDispatch(e)(dispatch)(buildAddFeatureAction)([newPositivePhones, newNegativePhones, feature])} onClick={e => props.dispatch(buildReducerAction(e, newPositivePhones, newNegativePhones, feature))}
value="Add feature" value="Add feature"
></input> ></input>
</form> </form>

View file

@ -1,42 +0,0 @@
div.Features {
ul.Features__list {
width: 100%;
li {
display: grid;
gap: 0.5em;
grid-template-columns: 10fr 10fr 1fr;
margin: 0.5em 0;
place-items: center center;
span.feature--names-and-phones {
display: grid;
grid-template-columns: repeat(auto-fit, minmax(100px, 1fr));
place-items: center center;
}
span.feature-name {
font-weight: 600;
}
}
}
form {
display: flex;
flex-flow: column;
input {
margin: 0.1em;
font-size: 1em;
}
}
button.delete-feature {
background-color: red;
border-color: transparent;
border-radius: 0.5em;
color: white;
max-height: 1.5em;
}
}

View file

@ -19,4 +19,19 @@ describe('Features', () => {
expect(getByTestId('Features')).toHaveTextContent('Phonetic Features'); expect(getByTestId('Features')).toHaveTextContent('Phonetic Features');
}); });
it('renders features from phonemes hook', () => {
const nPhone = {n:{
grapheme: 'n',
features: { nasal: true, occlusive: true, vowel: false } }}
const { getByTestId } = render(<Features phones={{nPhone}}
features={{
nasal: {positive: [nPhone.n], negative: []},
occlusive:{ positive: [nPhone.n], negative:[]},
vowel:{positive: [], negative: [nPhone.n]}
}}
/>);
expect(getByTestId('Features-list'))
.toContainHTML('<ul class="Features__list" data-testid="Features-list"><li><span class="plus-phones">[+ nasal] = n</span><span class="minus-phones">[- nasal] = </span></li><li><span class="plus-phones">[+ occlusive] = n</span><span class="minus-phones">[- occlusive] = </span></li><li><span class="plus-phones">[+ vowel] = </span><span class="minus-phones">[- vowel] = n</span></li></ul>');
});
}); });

View file

@ -1,28 +0,0 @@
import React from 'react';
import './Latl.scss';
const Latl = ({latl, dispatch}) => {
const { innerWidth, innerHeight } = window;
const handleChange = e => {
const setLatlAction = {
type: 'SET_LATL',
value: e.target.value
}
dispatch(setLatlAction);
}
return (
<div className="Latl">
<h3>.LATL</h3>
<textarea name="latl" id="latl"
value={latl}
cols={'' + Math.floor(innerWidth / 15)}
rows={'' + Math.floor(innerHeight / 30)}
onChange={handleChange}
/>
</div>
);
}
export default Latl;

View file

@ -1,3 +0,0 @@
div.Latl {
min-width: 80vw;
}

View file

@ -1,69 +0,0 @@
import React from 'react';
import './LatlOutput.scss';
import Output from './Output';
const LatlOutput = ({results, options, dispatch, errors, parseResults}) => {
const handleClick = e => dispatchFunc => {
e.preventDefault()
return dispatchFunc();
}
const dispatchClear = () => {
const clearAction = {
type: 'CLEAR',
value: {}
}
dispatch(clearAction)
}
const dispatchParse = () => {
const parseAction = {
type: 'PARSE_LATL',
value: {}
}
dispatch(parseAction)
}
const dispatchRun = () => {
const runAction = {
type: 'RUN',
value: {}
}
dispatch(runAction)
}
return (
<div className="LatlOutput">
<h3>Output</h3>
<form>
<input
className="form form--remove"
type="submit"
onClick={e=>handleClick(e)(dispatchClear)}
value="Clear"
/>
<input
id="Parse"
name="Parse"
className="form form--add"
type="submit"
onClick={e=>handleClick(e)(dispatchParse)}
value="Parse"
/>
<input
id="Run"
name="Run"
className="form form--add"
type="submit"
onClick={e=>handleClick(e)(dispatchRun)}
value="Run"
/>
</form>
<Output results={results} errors={errors} options={options} parseResults={parseResults}/>
</div>
);
}
export default LatlOutput;

View file

@ -1,9 +0,0 @@
div.LatlOutput {
display: flex;
flex-flow: column nowrap;
form {
display: grid;
grid-template-columns: repeat(auto-fit, min-max(10em, 1fr));
}
}

View file

@ -2,54 +2,57 @@ import React, { useState } from 'react';
import './Options.scss'; import './Options.scss';
import ls from 'local-storage'; import ls from 'local-storage';
const Options = ({ options, dispatch }) => { const Options = props => {
const [ load, setLoad ] = useState(''); const [ load, setLoad ] = useState('');
const handleRadioChange = e => { const handleRadioChange = e => {
const { name, id } = e.target; const option = e.target.name;
dispatch({ const setValue = e.target.id;
props.dispatch({
type: 'SET_OPTIONS', type: 'SET_OPTIONS',
value: { value: {
option: name, option,
setValue: id setValue
} }
}); });
} }
const handleFormSubmit = (e, options) => { const handleCheckChange = e => {
e.preventDefault(); const option = e.target.name;
dispatch({ const setValue = e.target.checked ? 'true' : 'false';
type: 'RUN', props.dispatch({
value: options type: 'SET_OPTIONS',
value: {
option,
setValue
}
}); });
} }
const handleOutputClearSubmit = e => { const handleFormSubmit = e => {
e.preventDefault(); e.preventDefault();
console.log('clearing')
dispatch({
type: 'CLEAR',
value: {}
});
} }
return ( return (
<div className="Options" data-testid="Options"> <div className="Options" data-testid="Options">
<h3>Modeling Options</h3> <h3>Modeling Options</h3>
<form onSubmit={e=>handleFormSubmit(e, options)} data-testid="Options-form"> <form onSubmit={e=>handleFormSubmit(e)} data-testid="Options-form">
{/* <h5>Output</h5> */}
<input <input
type="radio" name="output" id="default" type="radio" name="output" id="default"
checked={options ? options.output === 'default' : true} checked={props.options ? props.options.output === 'default' : true}
onChange={e=>handleRadioChange(e)} onChange={e=>handleRadioChange(e)}
/> />
<label htmlFor="default">Default <label htmlFor="default">Default
<span className="Options__output-example"> output</span> <span className="Options__output-example"> output</span>
</label> </label>
{/* <input <input
type="radio" name="output" id="proto" type="radio" name="output" id="proto"
checked={options ? options.output === 'proto' : false} checked={props.options ? props.options.output === 'proto' : false}
onChange={e=>handleRadioChange(e)} onChange={e=>handleRadioChange(e)}
/> />
<label htmlFor="proto">Proto <label htmlFor="proto">Proto
@ -58,19 +61,25 @@ const Options = ({ options, dispatch }) => {
<input <input
type="radio" name="output" id="diachronic" type="radio" name="output" id="diachronic"
checked={options ? options.output === 'diachronic' : false} checked={props.options ? props.options.output === 'diachronic' : false}
onChange={e=>handleRadioChange(e)} onChange={e=>handleRadioChange(e)}
/> />
<label htmlFor="diachronic">Diachronic <label htmlFor="diachronic">Diachronic
<span className="Options__output-example"> *proto > *epoch > output</span> <span className="Options__output-example"> *proto > *epoch > output</span>
</label> */} </label>
<input className="form form--add" type="submit" value="Run Changes"></input> <input
<input className="form form--remove" type="button" value="Clear Output" onClick={e=>handleOutputClearSubmit(e)}/> type="checkbox" name="save"
checked={props.options ? props.options.save : false}
onChange={e=>handleCheckChange(e)}
/>
<label htmlFor="save">Store session on Run</label>
<input type="submit" value="Run Changes"></input>
</form> </form>
{/* <form onSubmit={()=>{}}> <form onSubmit={() => {}}>
<label> <label>
Load from a prior run: Load from a prior run:
<select value={load} onChange={e=>setLoad(e.target.value)}> <select value={load} onChange={e=>setLoad(e.target.value)}>
@ -82,7 +91,7 @@ const Options = ({ options, dispatch }) => {
</select> </select>
</label> </label>
<input type="submit" value="Submit" /> <input type="submit" value="Submit" />
</form> */} </form>
</div> </div>
); );
} }

View file

@ -1,9 +0,0 @@
div.Options {
form {
display: grid;
grid-template-columns: 1fr 1fr;
gap: 0.5em;
}
}

View file

@ -19,4 +19,9 @@ describe('Options', () => {
expect(getByTestId('Options')).toHaveTextContent('Modeling Options'); expect(getByTestId('Options')).toHaveTextContent('Modeling Options');
}); });
it('renders form options from props', () => {
let options = {output: 'proto', save: true}
const { getByTestId } = render(<Options options={options} />)
expect(getByTestId('Options-form')).toHaveFormValues(options);
})
}); });

View file

@ -2,34 +2,12 @@ import React from 'react';
import './Output.scss'; import './Output.scss';
const Output = props => { const Output = props => {
const { results, options, errors, parseResults } = props;
const renderResults = () => {
switch(options.output) {
case 'default':
return renderDefault();
default:
return <></>
}
}
const renderDefault = () => {
return results.map((epoch, i) => {
const lexicon = epoch.lexicon.map((lexeme, i) => <span key={`${epoch.pass}-${i}`}>{lexeme}</span>);
return (
<div key={`epoch-${i}`} className="Output-epoch">
<h5>{epoch.pass}</h5>
<p className="lexicon">{lexicon}</p>
</div>
)
})
}
return ( return (
<div className="Output" data-testid="Output"> <div className="Output" data-testid="Output">
<h3>Results of Run</h3> <h3>Results of Run</h3>
<div data-testid="Output-lexicon" className="Output__container">
{parseResults ? parseResults : <></>} <div data-testid="Output-lexicon">
{results && results.length ? renderResults() : <></>} {props.results ? props.results.map((lexicalItem, i) => <p key={`output-lexical-item-${i}`}>{lexicalItem}</p>) : <></>}
</div> </div>
</div> </div>
); );

View file

@ -1,18 +0,0 @@
div.Output {
div.Output__container {
display: flex;
flex-flow: row wrap;
}
div.Output-epoch {
display: flex;
flex-flow: column;
p.lexicon {
display: grid;
grid-template-columns: repeat(auto-fit, minmax(5em, 1fr));
}
}
}

View file

@ -20,10 +20,8 @@ describe('Output', () => {
}); });
it('renders output lexicon list from output hook', () => { it('renders output lexicon list from output hook', () => {
const { getByTestId } = render(<Output results={[{pass: 'test', lexicon: ['word', 'lex', 'word']}]} options={{output: 'default'}}/>); const { getByTestId } = render(<Output results={['word', 'lex', 'word']}/>);
expect(getByTestId('Output-lexicon')).toContainHTML(wordListWordHTML); expect(getByTestId('Output-lexicon')).toContainHTML('<p>word</p><p>lex</p><p>word</p>');
}); });
}); });
const wordListWordHTML = '<div class="Output-epoch"><h5>test</h5><p class="lexicon"><span>word</span><span>lex</span><span>word</span></p></div>';

View file

@ -1,26 +1,7 @@
import React from 'react'; import React from 'react';
import './ProtoLang.scss'; import './ProtoLang.scss';
const ProtoLang = ({ lexicon, dispatch }) => { const ProtoLang = (props) => {
const getProperty = property => object => object[property];
const renderLexicon = () => {
if (!lexicon) return '';
// Code for optionally rendering epoch name with lexeme
// `\t#${lexeme.epoch.name}`
return lexicon.map(getProperty('lexeme')).join('\n');
}
const handleChange = e => {
const value = e.target.value.split(/\n/).map(line => {
const lexeme = line.split('#')[0].trim();
const epoch = line.split('#')[1] || '';
return { lexeme, epoch }
})
dispatch({
type: 'SET_LEXICON',
value
})
}
return ( return (
<div className="ProtoLang" data-testid="ProtoLang"> <div className="ProtoLang" data-testid="ProtoLang">
<h3>Proto Language Lexicon</h3> <h3>Proto Language Lexicon</h3>
@ -28,12 +9,24 @@ const ProtoLang = ({ lexicon, dispatch }) => {
<form data-testid="ProtoLang-Lexicon"> <form data-testid="ProtoLang-Lexicon">
<textarea <textarea
name="lexicon" name="lexicon"
cols="30"
rows="10"
data-testid="ProtoLang-Lexicon__textarea" data-testid="ProtoLang-Lexicon__textarea"
value={renderLexicon()} value={props.lexicon ? props.lexicon.map(lexeme => `${lexeme.lexeme} \t#${lexeme.epoch.name}`).join('\n'): ''}
onChange={e => handleChange(e)} onChange={e=> {
> console.log(e.target.value.split(/\n/).map(line => {
const lexeme = line.split('#')[0].trim();
const epoch = line.split('#')[1] || '';
return { lexeme, epoch }
}))
props.dispatch({
type: 'SET_LEXICON',
value: e.target.value.split(/\n/).map(line => {
const lexeme = line.split('#')[0].trim();
const epoch = line.split('#')[1] || '';
return { lexeme, epoch }
})
})
}
}>
</textarea> </textarea>
</form> </form>
</div> </div>

View file

@ -0,0 +1,5 @@
.ProtoLang {
width: 100%;
height: 100%;
color: black;
}

View file

@ -21,7 +21,7 @@ describe('ProtoLang', () => {
it('renders lexicon from state', () => { it('renders lexicon from state', () => {
const { getByTestId } = render(<ProtoLang lexicon={[{ lexeme:'one', epoch:{name: 'epoch-one', changes: []} }]}/>); const { getByTestId } = render(<ProtoLang lexicon={[{ lexeme:'one', epoch:{name: 'epoch-one', changes: []} }]}/>);
expect(getByTestId('ProtoLang-Lexicon')).toHaveFormValues({lexicon: 'one'}); expect(getByTestId('ProtoLang-Lexicon')).toHaveFormValues({lexicon: 'one \t#epoch-one'});
}); });
}) })

View file

@ -1,90 +1,27 @@
import React, { useState, useEffect } from 'react'; import React, { useState } from 'react';
import './SoundChangeSuite.scss'; import './SoundChangeSuite.scss';
const SoundChangeSuite = props => { const SoundChangeSuite = props => {
const { epochIndex, error, removeEpoch, epochs } = props; const [ epoch, setEpoch ] = useState(props.epoch ? props.epoch : {name:'', changes:['']});
const [ epoch, setEpoch ] = useState(props.epoch ? props.epoch : {name:'', changes:[''], parent:'none'});
const changeHandler = (e,cb) => { const changeHandler = (e,cb) => {
cb(e); cb(e);
props.updateEpoch(epoch, epochIndex); props.updateEpoch(epoch, props.epochIndex);
}
useEffect(() => {
props.updateEpoch(epoch, epochIndex);
}, [epoch])
const renderOptionFromEpoch = thisEpoch => (
<option
key={`${epoch.name}__parent-option--${thisEpoch.name}`}
value={thisEpoch.name}
>
{thisEpoch.name}
</option>
)
const replaceCurrentEpoch = thisEpoch => {
if (thisEpoch.name === epoch.name) return {name: 'none'}
return thisEpoch;
}
const isViableParent = thisEpoch => {
if (thisEpoch.parent && thisEpoch.parent === epoch.name) return false;
return true;
}
const parentsOptions = () => {
return epochs.map(replaceCurrentEpoch).filter(isViableParent).map(renderOptionFromEpoch)
}
const renderParentInput = () => {
if (epochIndex) return (
<>
<label htmlFor={`${epoch.name}-parent`}>
Parent Epoch:
</label>
<select
name="parent"
list={`${epoch.name}-parents-list`}
value={epoch.parent || 'none'}
onChange={e=>changeHandler(
e, ()=>setEpoch({...epoch, parent:e.target.value})
)
}
>
{parentsOptions()}
</select>
</>
)
return <></>
}
const renderError = () => {
if (error) return (
<p className="error">{error}</p>
)
return <></>
} }
return ( return (
<> <div className="SoundChangeSuite" data-testid={`${epoch.name}_SoundChangeSuite`}>
<h4>{epoch.name}</h4> <h4>{epoch.name}</h4>
{renderError()}
<form className="SoundChangeSuite__form" data-testid={`${epoch.name}_SoundChangeSuite_changes`}> <form className="SoundChangeSuite__form" data-testid={`${epoch.name}_SoundChangeSuite_changes`}>
<label htmlFor={`${epoch.name}-name`}>
Name: <textarea
</label>
<input type="text"
name="epoch" name="epoch"
id={`${epoch.name}-name`} cols="30" rows="1" id="" cols="30" rows="1"
value={epoch.name} value={epoch.name}
onChange={e=>changeHandler( onChange={e=>changeHandler(
e, () => { e, () => setEpoch({...epoch, name:e.target.value})
setEpoch({...epoch, name:e.target.value})
}
)} )}
></input> ></textarea>
{renderParentInput()}
<textarea <textarea
name="changes" name="changes"
@ -100,10 +37,10 @@ const SoundChangeSuite = props => {
)} )}
></textarea> ></textarea>
</form> </form>
<form onSubmit={e=>removeEpoch(e, epoch.name)}> <form onSubmit={e=>props.removeEpoch(e, epoch.name)}>
<input className="form form--remove" type="submit" name="remove-epoch" value={`remove ${epoch.name}`}></input> <input type="submit" name="remove-epoch" value={`remove ${epoch.name}`}></input>
</form> </form>
</> </div>
); );
} }

View file

@ -8,18 +8,18 @@ import extendExpect from '@testing-library/jest-dom/extend-expect'
it('renders SoundChangeSuite without crashing', () => { it('renders SoundChangeSuite without crashing', () => {
const div = document.createElement('div'); const div = document.createElement('div');
ReactDOM.render(<SoundChangeSuite epoch={{name:'Epoch Name', changes:['sound change rule']}} updateEpoch={()=>{}} removeEpoch={()=>{}}/>, div); ReactDOM.render(<SoundChangeSuite />, div);
ReactDOM.unmountComponentAtNode(div); ReactDOM.unmountComponentAtNode(div);
}); });
describe('SoundChangeSuite', () => { describe('SoundChangeSuite', () => {
it('renders the correct subtitle', () => {
const { getByTestId } = render(<SoundChangeSuite epoch={{name:'Epoch Name', changes:['sound change rule']}}/>);
expect(getByTestId('Epoch Name_SoundChangeSuite')).toHaveTextContent('Epoch Name');
});
it('renders a suite of soundchanges', () => { it('renders a suite of soundchanges', () => {
const { getByTestId } = render( const { getByTestId } = render(<SoundChangeSuite epoch={{name:'Epoch Name', changes:['sound>change/environment']}}/>);
<SoundChangeSuite epoch={{name:'Epoch Name', changes:['sound>change/environment']}}
updateEpoch={()=>{}} removeEpoch={()=>{}}
/>
);
expect(getByTestId('Epoch Name_SoundChangeSuite_changes')).toHaveFormValues({changes: 'sound>change/environment'}) expect(getByTestId('Epoch Name_SoundChangeSuite_changes')).toHaveFormValues({changes: 'sound>change/environment'})
}) })
}); });

15
src/index.css Normal file
View file

@ -0,0 +1,15 @@
body {
margin: 0;
font-family: -apple-system, BlinkMacSystemFont, "Segoe UI", "Roboto", "Oxygen",
"Ubuntu", "Cantarell", "Fira Sans", "Droid Sans", "Helvetica Neue",
sans-serif;
-webkit-font-smoothing: antialiased;
-moz-osx-font-smoothing: grayscale;
background-color: #111;
color: #EEE;
}
code {
font-family: source-code-pro, Menlo, Monaco, Consolas, "Courier New",
monospace;
}

View file

@ -1,12 +1,10 @@
import React from 'react'; import React from 'react';
import ReactDOM from 'react-dom'; import ReactDOM from 'react-dom';
import { HashRouter as Router } from 'react-router-dom'; import './index.css';
import './index.scss';
import App from './App'; import App from './App';
import * as serviceWorker from './serviceWorker'; import * as serviceWorker from './serviceWorker';
ReactDOM.render(<Router><App /></Router> ReactDOM.render(<App />, document.getElementById('root'));
, document.getElementById('root'));
// If you want your app to work offline and load faster, you can change // If you want your app to work offline and load faster, you can change
// unregister() to register() below. Note this comes with some pitfalls. // unregister() to register() below. Note this comes with some pitfalls.

View file

@ -1,24 +0,0 @@
@import '../public/stylesheets/variables';
body {
margin: 0;
font-family: 'Catamaran', Arial, Helvetica, sans-serif;
background-color: map-get($colors, 'main--bg');
color: map-get($colors, 'main');
textarea, input[type="text"] {
background-color: map-get($colors, 'text-input--bg');
color: map-get($colors, 'text-input');
border: 1px solid map-get($colors, 'main');
font-family: 'Fira Code', monospace;
}
code {
font-family: 'Fira Code', monospace;
}
p.error {
color: map-get($colors, 'error');
}
}

View file

@ -1,3 +0,0 @@
export const clearOutput = (state, action) => {
return { ...state, results: [], errors: {}, parseResults: '' };
}

View file

@ -1,739 +0,0 @@
// @flow
import type { stateType } from './reducer';
export type initAction = {
type: "INIT"
}
export const clearState = () => {
return {
epochs: [],
phones: {},
options: { output: 'default', save: false },
results: [],
errors: {},
features: {},
lexicon: [],
latl: '',
parseResults: ''
}
}
export const waffleState = () => {
return {
epochs: [],
phones: {},
options: { output: 'default', save: false },
results: [],
errors: {},
features: {},
lexicon: [],
latl: waffleLatl,
parseResults: ''
}
}
export const initState = (changesArgument: number): stateType => {
const state = {
epochs: [
{
name: 'epoch-1',
changes: [
'[+ occlusive - nasal]>[+ occlusive + nasal]/n_.',
'a>ɯ/._#',
'[+ sonorant - low rounded high back]>0/._.',
'[+ obstruent]>[+ obstruent aspirated ]/#_.',
'[+ sonorant - rounded]>[+ sonorant + rounded]/._#',
// 'at>ta/._#'
]
}
],
phones: {
a: {
grapheme: 'a', features: {
sonorant: true, back: true, low: true, high: false, rounded: false
}
},
u: {
grapheme: 'u', features: {
sonorant: true, back: true, low: false, high: true, rounded: true,
}
},
ɯ: {
grapheme: 'ɯ', features: {
sonorant: true, back: true, low: false, high: true, rounded: false,
}
},
ə: {
grapheme: 'ə', features: {
sonorant: true, low: false, rounded: false, high: false, back: false
}
},
t: {
grapheme: 't', features: {
occlusive: true, coronal: true, obstruent: true, nasal: false
},
ʰ: {
grapheme: 'tʰ', features: {
occlusive: true, coronal: true, obstruent: true, aspirated: true
}
}
},
n: {
grapheme: 'n', features: {
sonorant: true, nasal: true, occlusive: true, coronal: true
}
}
},
options: {
output: 'default', save: false
},
results: [],
errors: {},
features: {},
lexicon: [],
latl: '',
parseResults: ''
};
state.features = {
sonorant: { positive:[ state.phones.a, state.phones.u, state.phones.ɯ, state.phones.ə, state.phones.n], negative: [] },
back: { positive:[ state.phones.a, state.phones.u, state.phones.ɯ ], negative: [ state.phones.ə ] },
low: { positive:[ state.phones.a ], negative: [ state.phones.u, state.phones.ɯ, state.phones.ə ] },
high: { positive:[ state.phones.u, state.phones.ɯ ], negative: [ state.phones.a, state.phones.ə ] },
rounded: { positive:[ state.phones.u ], negative: [ state.phones.a, state.phones.ɯ, state.phones.ə ] },
occlusive: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
coronal: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
obstruent: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
nasal: { positive:[ state.phones.n ], negative: [state.phones.t, state.phones.t.ʰ] },
aspirated: { positive:[ state.phones.t.ʰ ], negative: [ state.phones.t ] },
}
state.lexicon = [
{lexeme: 'anta', epoch: state.epochs[0]},
{lexeme: 'anat', epoch: state.epochs[0]},
{lexeme: 'anət', epoch: state.epochs[0]},
{lexeme: 'anna', epoch: state.epochs[0]},
{lexeme: 'tan', epoch: state.epochs[0]},
{lexeme: 'ənta', epoch: state.epochs[0]}
]
if(changesArgument > -1) state.epochs[0].changes = state.epochs[0].changes.splice(0, changesArgument)
return state;
}
const waffleLatl = `
; -------- main class features
[consonantal
+=
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; AFFRICATES
/ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; NASALS
m ɱ / n / ŋ /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; GLIDES
j / w /
; LARYNGEALS
h ɦ / ʔ
]
[sonorant
+=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; GLIDES
j / w w̥ /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; NASALS
m ɱ / n / ŋ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
-=
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; AFFRICATES
/ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; LARYNGEALS
h ɦ / ʔ
]
[approximant
+=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; GLIDES
j / w /
; SYLLABIC LIQUIDS
l̩ / ɹ̩
-=
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; AFFRICATES
/ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; NASALS
m ɱ / n / ŋ /
; SYLLABIC NASALS
m̩ / n̩
]
; -------- laryngeal features
[voice
+=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; GLIDES
j / w /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; NASALS
m ɱ / n / ŋ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩ /
; VOICED FRICATIVES
v / ð / z / ʒ /
; VOICED AFFRICATES
/
; VOICED LARYNGEALS
; LARYNGEALS
ɦ
-= voiceless obstruents
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; VOICELESS AFFRICATES
/ /
; VOICELESS FRICATIVES
f / θ / s / ʃ / ç / x /
; VOICELESS LARYNGEALS
h / ʔ
]
[spreadGlottis
+=
; ASPIRATED PLOSIVES
/ / /
; ASPIRATED AFFRICATES
/
; SPREAD LARYNGEALS
h ɦ
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; UNASPIRATED PLOSIVES
p / pʼ / t / tʼ / ɾ / k / kʼ /
; UNASPIRATED AFFRICATES
/ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; NASAL OBSTRUENTS
m ɱ / n / ŋ /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩ /
; GLIDES
j / w
; CONSTRICTED LARYNGEALS
ʔ
]
[constrictedGlottis
+=
; LARYNGEALIZED RHOTIC
ɹˤ /
; CONSTRICTED LARYNGEAL
ʔ /
; EJECTIVE PLOSIVES
pʼ / tʼ / kʼ
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; PLOSIVES
p / / t / ɾ / k / /
; AFFRICATES
/ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; NASAL OBSTRUENTS
m ɱ / n / ŋ /
; LIQUIDS
l /
; NON-PHARYNGEALIZED RHOTICS
ɹ ɹʲ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; GLIDES
j / w
; SPREAD LARYNGEALS
h ɦ /
]
; -------- manner features
[continuant
+=
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; GLIDES
j / w /
; SYLLABIC LIQUIDS
l̩ / ɹ̩ /
; TAPS
ɾ
-=
; NON-TAP PLOSIVES
p / pʼ / / t / tʼ / / k / kʼ / /
; AFFRICATES
/ /
; NASALS
m ɱ / n / ŋ /
; SYLLABIC NASALS
m̩ / n̩
]
[nasal
+=
; NASALS
m ɱ / n / ŋ /
; SYLLABIC NASALS
m̩ / n̩
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x /
; LIQUIDS + RHOTICS
l / ɹ ɹʲ ɹˤ /
; GLIDES
j / w /
; SYLLABIC LIQUIDS
l̩ / ɹ̩ /
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; AFFRICATES
/ /
]
[strident
+=
; STRIDENT FRICATIVES
f / v / s / z / ʃ / ʒ /
; STRIDENT AFFRICATES
/
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ / /
; NON-STRIDENT FRICATIVES
θ / ð / ç / x /
; NASAL OBSTRUENTS
m ɱ / n / ŋ /
; RHOTICS + LIQUIDS
l / ɹ ɹʲ ɹˤ /
; SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩ /
; GLIDES
j / w
]
[lateral
+=
; LATERAL LIQUIDS
l /
; SYLLABIC LATERALS /
l̩
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; PLOSIVES
p / pʼ / / t / tʼ / ɾ / k / kʼ /
; AFFRICATES
/
; FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ / ç / x
; NASAL OBSTRUENTS
m ɱ / n / ŋ
; RHOTIC LIQUIDS
ɹ ɹʲ ɹˤ
; NON-LIQUID SYLLABIC CONSONANTS
m̩ / n̩ / ɹ̩
; GLIDES
j / w
]
; -------- ---- PLACE features
; -------- labial features
[labial
+=
; ROUNDED VOWELS
u̟ / ʊ̞ / ɔ /
; LABIAL PLOSIVES
p / pʼ / /
; LABIAL FRICATIVES
f / v /
; LABIAL NASALS
m ɱ /
; LABIAL SYLLABIC CONSONANTS
m̩ /
; LABIAL GLIDES
w
-=
; UNROUNDED VOWELS
æ / e / ə / ɑ / ɪ̞ / ɛ / ʌ / i /
; NON-LABIAL PLOSIVES
t / tʼ / ɾ / k / kʼ / /
; NON-LABIAL AFFRICATES
/ /
; NON-LABIAL FRICATIVES
θ / ð / s / z / ʃ / ʒ / ç / x /
; NON-LABIAL NASAL OBSTRUENTS
n / ŋ /
; LIQUIDS
l /
; RHOTIC LIQUIDS
ɹ ɹʲ ɹˤ /
; NON-LABIAL SYLLABIC CONSONANTS
n̩ / l̩ / ɹ̩ /
; NON-LABIAL GLIDES
j
]
; -------- coronal features
[coronal
+=
; CORONAL PLOSIVES
t / tʼ / ɾ /
; CORONAL AFFRICATES
/ /
; CORONAL FRICATIVES
θ / ð / s / z / ʃ / ʒ /
; CORONAL NASALS
n /
; CORONAL LIQUIDS
l
; CORONAL RHOTIC LIQUIDS
ɹ
; CORONAL SYLLABIC CONSONANTS
n̩ / l̩ / ɹ̩
-=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; NON-CORONAL PLOSIVES
p / pʼ / / k / kʼ /
; NON-CORONAL FRICATIVES
f / v / ç / x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ / ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩ /
; NON-CORONAL GLIDES
j / w
]
[anterior
+=
; ALVEOLAR PLOSIVES
t / tʼ / ɾ /
; ALVEOLAR AFFRICATES
/ /
; DENTAL FRICATIVES
θ / ð /
; ALVEOLAR FRICATIVES
s / z /
; ALVEOLAR NASALS
n /
; ALVEOLAR LIQUIDS
l
; ALVEOLAR SYLLABIC CONSONANTS
n̩ / l̩ /
-=
; POSTALVEOLAR FRICATIVES
ʃ / ʒ /
; POSTALVEOLAR RHOTIC LIQUIDS
ɹ /
; POSTALVEOLAR SYLLABIC CONSONANTS
ɹ̩ /
; -- NON-CORONALs
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; NON-CORONAL PLOSIVES
p / pʼ / / k / kʼ /
; NON-CORONAL FRICATIVES
f / v / ç / x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ / ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩ /
; NON-CORONAL GLIDES
j / w
]
[distributed
+=
; DENTAL FRICATIVES
θ / ð /
; POSTALVEOLAR FRICATIVES
ʃ / ʒ /
; POSTALVEOLAR RHOTIC LIQUIDS
ɹ /
; POSTALVEOLAR SYLLABIC CONSONANTS
ɹ̩ /
-=
; apical / retroflex
; ALVEOLAR PLOSIVES
t / tʼ / ɾ /
; ALVEOLAR FRICATIVES
s / z /
; ALVEOLAR NASALS
n /
; ALVEOLAR LIQUIDS
l
; ALVEOLAR SYLLABIC CONSONANTS
n̩ / l̩ /
; -- NON-CORONALS
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; NON-CORONAL PLOSIVES
p / pʼ / / k / kʼ /
; NON-CORONAL FRICATIVES
f / v / ç / x
; NON-CORONAL NASAL OBSTRUENTS
m ɱ / ŋ
; NON-CORONAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; NON-CORONAL SYLLABIC CONSONANTS
m̩ /
; NON-CORONAL GLIDES
j / w
]
; -------- dorsal features
[dorsal
+=
; VOWELS
æ / e / ə / ɑ / ɔ / ɪ̞ / ɛ / ʌ / ʊ̞ / i / u̟ /
; DORSAL PLOSIVES
k / kʼ / /
; DORSAL FRICATIVES
ç / x /
; DORSAL NASAL OBSTRUENTS
ŋ /
; DORSAL RHOTIC LIQUIDS
ɹʲ ɹˤ
; DORSAL GLIDES
j
-=
; NON-DORSAL PLOSIVES
p / pʼ / / t / tʼ / ɾ /
; NON-DORSAL AFFRICATES
/ /
; NON-DORSAL FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ /
; NON-DORSAL NASALS
m ɱ / n /
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; NON-DORSAL GLIDES
w
]
[high
+=
; HIGH VOWELS
i / u̟ / ʊ̞ / ɪ̞
; HIGH DORSAL PLOSIVES
k / kʼ / /
; HIGH DORSAL FRICATIVES
ç / x /
; HIGH DORSAL NASAL OBSTRUENTS
ŋ /
; HIGH RHOTIC LIQUIDS
ɹʲ
; HIGH DORSAL GLIDES
j / w
-= χ / e / o / a
; NON-HIGH VOWELS
ɑ / æ / e / ə / ɛ / ʌ
; NON-HIGH RHOTIC LIQUIDS
ɹˤ
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p / pʼ / / t / tʼ / ɾ /
; NON-DORSAL AFFRICATES
/ /
; NON-DORSAL FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ /
; NON-DORSAL NASALS
m ɱ / n /
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; NON-DORSAL GLIDES
w
]
[low
+=
; LOW VOWELS
ɑ / æ / ɛ /
; LOW DORSAL RHOTIC LIQUIDS
ɹˤ
-= a / ɛ / ɔ
; NON-LOW VOWELS
i / u̟ / ʊ̞ / ɪ̞ / e / ə / ʌ
; NON-LOW DORSAL PLOSIVES
k / kʼ / /
; NON-LOW DORSAL FRICATIVES
ç / x /
; NON-LOW DORSAL NASAL OBSTRUENTS
ŋ /
; NON-LOW DORSAL RHOTIC LIQUIDS
ɹʲ
; DORSAL GLIDES
j
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p / pʼ / / t / tʼ / ɾ /
; NON-DORSAL AFFRICATES
/ /
; NON-DORSAL FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ /
; NON-DORSAL NASALS
m ɱ / n /
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; NON-DORSAL GLIDES
w
]
[back
+=
; BACK VOWELS
ɑ / ɔ / ʌ / ʊ̞ / u̟ /
; BACK DORSAL PLOSIVES
k / kʼ / /
; BACK DORSAL FRICATIVES
x /
; BACK DORSAL NASAL OBSTRUENTS
ŋ /
; BACK DORSAL RHOTIC LIQUIDS
ɹˤ
-=
; NON-BACK DORSAL FRICATIVES
ç /
; NON-BACK DORSAL RHOTIC LIQUIDS
ɹʲ
; NON-BACK DORSAL GLIDES
j
; NON-BACK VOWELS
æ / e / ə / ɪ̞ / ɛ / i
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p / pʼ / / t / tʼ / ɾ /
; NON-DORSAL AFFRICATES
/ /
; NON-DORSAL FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ /
; NON-DORSAL NASALS
m ɱ / n /
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; NON-DORSAL GLIDES
w
]
[tense ; compare to ATR or RTR
+=
; TENSE VOWELS
e / i / u̟ / ɑ
-=
; NON-TENSE VOWELS
æ / ə / ɪ̞ / ɛ / ʌ / ʊ̞ / ɔ /
; DORSAL PLOSIVES
k / kʼ / /
; DORSAL FRICATIVES
ç / x /
; DORSAL NASAL OBSTRUENTS
ŋ /
; DORSAL RHOTIC LIQUIDS
ɹʲ ɹˤ /
; DORSAL GLIDES
j
; -- NON-DORSALS
; NON-DORSAL PLOSIVES
p / pʼ / / t / tʼ / ɾ /
; NON-DORSAL AFFRICATES
/ /
; NON-DORSAL FRICATIVES
f / v / θ / ð / s / z / ʃ / ʒ /
; NON-DORSAL NASALS
m ɱ / n /
; NON-DORSAL LIQUIDS
l
; NON-DORSAL RHOTIC LIQUIDS
ɹ
; NON-DORSAL SYLLABIC CONSONANTS
m̩ / n̩ / l̩ / ɹ̩
; NON-DORSAL GLIDES
w
]
*PROTO
; -- Devoicing, all our z's become s's
[+ voice - continuant]>[- voice]/._.
; -- Reduction of schwa
ə>0/._.
|Gif Lang
*PROTO
; -- Ejectivization, all our pits become pit's
[+ spreadGlottis - continuant]>[+ constrictedGlottis - spreadGlottis]/._[+ constrictedGlottis]
[+ spreadGlottis - continuant]>[+ constrictedGlottis - spreadGlottis]/[+ constrictedGlottis]_.
[+ constrictedGlottis]>0/[+ constrictedGlottis - continuant]_.
[+ constrictedGlottis]>0/._[+ constrictedGlottis - continuant]
|Jif Lang
`

View file

@ -1,529 +0,0 @@
import { stateReducer } from './reducer';
export const setLatl = (state, action) => {
let latl = action.value;
return {...state, latl, parseResults: ''};
}
const getOneToken = (latl, tokens) => {
for (const [type, regEx] of tokenTypes) {
const newRegEx = new RegExp(`^(${regEx})`);
const match = latl.match(newRegEx) || null;
if (match) {
const newTokens = [...tokens, {type, value: match[0].trim()}]
const newLatl = latl.slice(match[0].length ,);
return [newLatl, newTokens]
}
}
throw `Unexpected token at ${latl.split('\n')[0]}`
}
export const tokenize = latl => {
let i = 0;
let tokens = [];
let newLatl = latl.trim();
try {
while(newLatl.length) {
[newLatl, tokens] = getOneToken(newLatl, tokens)
}
return tokens;
}
catch (err) {
return {errors: 'tokenization error', message: err, newLatl}
}
}
const parseLineBreak = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
if (!lastNode) return tree;
switch (lastNode.type) {
case 'rule': {
if (tree[tree.length - 2].type === 'ruleSet') {
const ruleValue = lastNode.value;
tree[tree.length - 2].value.push(ruleValue);
tree.pop()
return tree;
}
if (tree[tree.length - 2].type === 'epoch') {
const newNode = { type: 'ruleSet', value: [ lastNode.value ] }
tree[tree.length - 1] = newNode;
return tree;
}
}
case 'feature--plus': {
// tree[tree.length - 1].type === 'feature';
return tree;
}
case 'feature--minus': {
// tree[tree.length - 1].type === 'feature';
return tree;
}
default:
return tree;
}
}
const parseWhiteSpace = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule': {
tree[tree.length - 1] = {...lastNode, value: lastNode.value + ' ' }
return tree;
}
default:
return tree;
}
}
const parseStar = (tree, token, index, tokens) => {
const nextToken = tokens[index + 1];
if (nextToken.type === 'referent') {
return [...tree, { type: 'epoch-parent' }]
}
}
const parsePipe = (tree, token, index, tokens) => {
const nextToken = tokens[index + 1];
if (nextToken.type === 'referent') {
const ruleToken = tree[tree.length - 1];
const epochToken = tree[tree.length - 2];
if (ruleToken.type === 'rule' || ruleToken.type === 'ruleSet') {
if (epochToken.type === 'epoch') {
tree[tree.length - 2] = {
...epochToken,
changes: [...ruleToken.value],
type: 'epoch-name'
}
tree.pop();
return tree;
}
}
}
return [...tree, 'unexpected pipe']
}
const parseReferent = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'epoch-parent': {
tree[tree.length - 1] = {...lastNode, parent: token.value, type: 'epoch' }
return tree;
}
case 'epoch-name': {
tree[tree.length - 1] = {...lastNode, name: token.value, type: 'epoch' }
return [...tree, { type: 'main'}];
}
case 'epoch': {
return [...tree, { type: 'rule', value: token.value } ]
}
case 'rule': {
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value }
return tree;
}
case 'ruleSet': {
return [...tree, { type: 'rule', value: token.value }]
}
case 'feature': {
if (!lastNode.value) {
tree[tree.length - 1].value = token.value;
return tree;
}
}
case 'feature--plus': {
if (lastNode.value) {
lastNode.positivePhones = [...lastNode.positivePhones, token.value ]
}
else {
lastNode.value = token.value;
}
tree[tree.length - 1] = lastNode;
return [...tree]
}
case 'feature--minus': {
if (lastNode.value) {
lastNode.negativePhones = [...lastNode.negativePhones, token.value ]
}
else {
lastNode.value = token.value;
}
tree[tree.length - 1] = lastNode;
return [...tree]
}
case 'lexicon': {
if (!lastNode.epoch) {
tree[tree.length - 1].epoch = token.value;
}
else {
tree[tree.length - 1].value.push(token.value)
}
return tree;
}
default:
return [...tree, `unexpected referent ${token.value}`]
}
}
const parsePhone = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch(lastNode.type) {
case 'rule': {
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value }
return tree;
}
case 'ruleSet': {
return [...tree, { type: 'rule', value: token.value }]
}
case 'feature--plus':
lastNode.positivePhones = [...lastNode.positivePhones, token.value ];
tree[tree.length - 1] = lastNode;
return tree;
case 'feature--minus':
lastNode.negativePhones = [...lastNode.negativePhones, token.value ];
tree[tree.length - 1] = lastNode;
return tree;
default:
return [...tree, `unexpected phone ${token.value}`]
}
}
const parseOpenBracket = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
if (lastNode) {
switch (lastNode.type) {
case 'epoch':
return [...tree, {type: 'rule', value: token.value}]
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value }
return tree;
case 'ruleSet':
return [...tree, {type: 'rule', value: token.value}];
// case 'feature':
// return [{type: 'feature', positivePhones: [], negativePhones: []}];
case 'feature--plus':
return [...tree, {type: 'feature', positivePhones: [], negativePhones: []}];
case 'feature--minus':
return [...tree, {type: 'feature', positivePhones: [], negativePhones: []}];
case 'main':
return [...tree, {type: 'feature', positivePhones: [], negativePhones: []}];
default:
return [...tree, 'unexpected open bracket']
}
}
return [{type: 'feature', positivePhones: [], negativePhones: []}]
}
const parseCloseBracket = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value }
return tree;
case 'feature--plus':
return tree;
case 'feature--minus':
return tree;
default:
return [...tree, 'unexpected close bracket']
}
}
const parsePositiveAssignment = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'feature':
tree[tree.length - 1].type = 'feature--plus'
return tree;
default:
return [...tree, 'unexpected positive assignment']
}
}
const parseNegativeAssignment = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'feature':
tree[tree.length - 1].type = 'feature--minus'
return tree;
case 'feature--plus':
tree[tree.length - 1].type = 'feature--minus';
return tree;
default:
return [...tree, 'unexpected negative assignment']
}
}
const parsePlus = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
case 'feature':
tree[tree.length - 1] = {...lastNode, type: 'feature--plus'}
return tree;
case 'feature--minus':
tree[tree.length - 1] = {...lastNode, type: 'feature--minus'}
return tree;
default:
return [...tree, 'unexpected plus']
}
}
const parseMinus = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
case 'feature':
tree[tree.length - 1] = {...lastNode, type: 'feature--minus'}
return tree;
default:
return [...tree, 'unexpected minus']
}
}
const parseEqual = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'feature--plus':
return tree;
case 'feature--minus':
return tree;
default:
return [...tree, 'unexpected equal'];
}
}
const parseGreaterThan = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
default:
return [...tree, 'unexpected greater than']
}
}
const parseSlash = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
if (lastNode) {
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
case 'feature--plus':
return tree;
case 'feature--minus':
return tree;
case 'lexicon':
return [...tree, { }];
case 'main':
return [...tree, { type: 'lexicon', value: []}]
default:
return [...tree, 'unexpected slash']
}
}
return [...tree, { type: 'lexicon', value: []}]
}
const parseHash = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
default:
return [...tree, 'unexpected hash']
}
}
const parseDot = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
default:
return [...tree, 'unexpected dot']
}
}
const parseUnderScore = (tree, token, index, tokens) => {
const lastNode = tree[tree.length - 1];
switch (lastNode.type) {
case 'rule':
tree[tree.length - 1] = {...lastNode, value: lastNode.value + token.value}
return tree;
default:
return [...tree, 'unexpected underscore']
}
}
const generateNode = (tree, token, index, tokens) => {
switch (token.type) {
// if comment, consume without effect
case 'semicolon':
return [...tree]
case 'lineBreak':
return parseLineBreak(tree, token, index, tokens);
case 'whiteSpace':
return parseWhiteSpace(tree, token, index, tokens);
// if *PROTO consume token:* and add epochs: [ { parent: 'PROTO' } ]
case 'star':
return parseStar(tree, token, index, tokens);
case 'pipe':
return parsePipe(tree, token, index, tokens);
case 'referent':
return parseReferent(tree, token, index, tokens);
case 'phone':
return parsePhone(tree, token, index, tokens);
case 'openBracket':
return parseOpenBracket(tree, token, index, tokens);
case 'closeBracket':
return parseCloseBracket(tree, token, index, tokens);
case 'positiveAssignment':
return parsePositiveAssignment(tree, token, index, tokens);
case 'negativeAssignment':
return parseNegativeAssignment(tree, token, index, tokens);
case 'plus':
return parsePlus(tree, token, index, tokens);
case 'minus':
return parseMinus(tree, token, index, tokens);
case 'equal':
return parseEqual(tree, token, index, tokens);
case 'greaterThan':
return parseGreaterThan(tree, token, index, tokens);
case 'slash':
return parseSlash(tree, token, index, tokens);
case 'hash':
return parseHash(tree, token, index, tokens);
case 'dot':
return parseDot(tree, token, index, tokens);
case 'underscore':
return parseUnderScore(tree, token, index, tokens);
default:
return [...tree, { ...token }]
}
}
const addToken = (tree, token, index, tokens) => generateNode(tree, token, index, tokens);
const connectNodes = (tree, node, index, nodes) => {
switch (node.type) {
case 'epoch':
delete node.type;
return {...tree, epochs: [...tree.epochs, {...node, index: tree.epochs.length } ] }
case 'feature':
node.feature = node.value;
delete node.value;
delete node.type;
return {...tree, features: [...tree.features, {...node } ] }
case 'feature--minus':
node.feature = node.value;
delete node.value;
delete node.type;
if (tree.features.length && tree.features[tree.features.length - 1].feature === node.feature) {
tree.features[tree.features.length - 1].negativePhones = node.negativePhones
return tree;
}
return {...tree, features: [...tree.features, {...node} ] }
case 'feature--plus':
delete node.type;
node.feature = node.value;
delete node.value;
if (tree.features.length && tree.features[tree.features.length - 1].feature === node.feature) {
tree.features[tree.features.length - 1].positivePhones = node.positivePhones
return tree;
}
return {...tree, features: [...tree.features, {...node} ] }
case 'lexicon':
delete node.type;
return {...tree, lexicon: [...tree.lexicon, node]}
default:
return tree;
}
}
export const buildTree = tokens => {
const bareTree = {
epochs: [],
features: [],
lexicon: []
}
const nodes = tokens.reduce(addToken, []);
// return nodes
const tree = nodes.reduce(connectNodes, bareTree);
const filterProps = Object.entries(tree).filter(([key, value]) => !value.length)
.map(([key, value]) => key)
return filterProps.reduce((tree, badProp) => {
delete tree[badProp];
return tree;
}, tree);
}
export const generateAST = latl => {
// tokenize
const tokens = tokenize(latl.trim());
// build tree
const tree = buildTree(tokens);
return tree;
}
export const parseLatl = (state, action) => {
try {
const latl = state.latl;
const AST = generateAST(latl);
const features = AST.features;
if (features) {
if (state.features) {
state = Object.keys(state.features).reduce((state, feature) => {
return stateReducer(state, {type: 'DELETE_FEATURE', value: feature})
}, state)
}
state = features.reduce((state, feature) => stateReducer(state, {type:'ADD_FEATURE', value: feature}), state);
}
delete AST.features;
const lexicon = AST.lexicon;
if (lexicon) {
if (state.lexicon) {
state.lexicon = [];
}
state = lexicon.reduce((state, epoch) => {
return epoch.value.reduce((reducedState, lexeme) => {
return stateReducer(reducedState, {type: 'ADD_LEXEME', value: { lexeme, epoch: epoch.epoch }})
}, state)
}, state)
}
delete AST.lexicon;
Object.entries(AST).forEach(([key, value]) => state[key] = value);
return { ...state, parseResults: 'latl parsed successfully', results:[] }
}
catch (e) {
console.log(e)
return { ...state, parseResults: 'error parsing', errors: e}
}
}
const tokenTypes = [
['semicolon', ';.*\n'],
[`star`, `\\*`],
['pipe', `\\|`],
['openBracket', `\\[`],
['closeBracket', `\\]`],
['positiveAssignment', `\\+=`],
['negativeAssignment', `\\-=`],
['plus', `\\+`],
['minus', `\\-`],
['greaterThan', `\\>`],
['hash', `#`],
['slash', `\/`],
['dot', `\\.`],
['underscore', `\\_`],
[`referent`, `[A-Za-z]+[\u00c0-\u03FFA-Za-z0-9\\-\\_]*`],
[`phone`, `[\u00c0-\u03FFA-Za-z0]+`],
['equal', `=`],
[`lineBreak`, `\\n`],
[`whiteSpace`, `\\s+`]
]

View file

@ -1,520 +0,0 @@
import { stateReducer } from './reducer';
import { initState } from './reducer.init';
import { tokenize, buildTree, parseLatl } from './reducer.latl';
describe('LATL', () => {
it('returns state unaltered with no action body', () => {
const state = initState();
const action = {
type: 'SET_LATL',
value: ''
}
const returnedState = stateReducer(state, action)
expect(returnedState).toStrictEqual(state);
})
it('returns tokens from well-formed latl epoch definition', () => {
const tokens = tokenize(epochDefinitionLatl);
expect(tokens).toStrictEqual(tokenizedEpoch)
});
it('returns tokens from well-formed latl feature definition', () => {
const tokens = tokenize(featureDefinitionLatl);
expect(tokens).toStrictEqual(tokenizedFeature);
});
it('returns tokens from well-formed latl lexicon definition', () => {
const tokens = tokenize(lexiconDefinitionLatl);
expect(tokens).toStrictEqual(tokenizedLexicon);
});
it('returns tokens from well-formed latl epoch, feature, and lexicon definitions', () => {
const latl = epochDefinitionLatl + '\n' + featureDefinitionLatl + '\n' + lexiconDefinitionLatl;
const tokens = tokenize(latl);
const lineBreaks = [{ type: 'lineBreak', value: '' },{ type: 'lineBreak', value: '' },{ type: 'lineBreak', value: '' }]
const tokenizedLatl = [...tokenizedEpoch, ...lineBreaks, ...tokenizedFeature, ...lineBreaks, ...tokenizedLexicon];
expect(tokens).toStrictEqual(tokenizedLatl);
});
it('returns AST from well-formed epoch tokens', () => {
const tree = buildTree(tokenizedEpoch);
expect(tree).toStrictEqual(treeEpoch);
})
it('returns AST from well-formed feature tokens', () => {
const tree = buildTree(tokenizedFeature);
expect(tree).toStrictEqual(treeFeature);
})
it('returns AST from well-formed lexicon tokens', () => {
const tree = buildTree(tokenizedLexicon);
expect(tree).toStrictEqual(treeLexicon);
})
it('parse returns state from well-formed feature latl', () => {
const state = initState();
const setAction = {
type: 'SET_LATL',
value: featureDefinitionLatl
}
const latlState = stateReducer(state, setAction);
const parseState = parseLatl(latlState, {});
expect(parseState).toStrictEqual(featureState)
})
it('returns run from well-formed epoch latl', () => {
const state = initState();
const setAction = {
type: 'SET_LATL',
value: runEpochLatl
}
const latlState = stateReducer(state, setAction);
const parseState = parseLatl(latlState, {})
// expect(parseState).toStrictEqual(epochState);
parseState.lexicon[0].epoch = 'PROTO'
const runState = stateReducer(parseState, {type: 'RUN', value:{}})
expect(runState).toStrictEqual({...runState, results: runEpochResults})
})
it('returns state from well-formed lexicon latl', () => {
const state = initState();
const setAction = {
type: 'SET_LATL',
value: lexiconDefinitionLatl
}
const latlState = stateReducer(state, setAction);
const parseState = parseLatl(latlState, {});
expect(parseState).toStrictEqual(lexiconState)
})
// it('returns state from well formed latl', () => {
// const state = initState();
// const setAction = {
// type: 'SET_LATL',
// value: totalLatl
// }
// const latlState = stateReducer(state, setAction);
// const parseState = parseLatl(latlState, {});
// expect(parseState).toStrictEqual(totalLatlState)
// })
})
const epochDefinitionLatl = `
; comment
*PROTO
[+ FEATURE]>[- FEATURE]/._.
n>m/#_.
|CHILD
`
const runEpochLatl = `
; comment
*PROTO
a>u/._.
|epoch-1
`
const runEpochResults = [
{
pass: 'epoch-1',
parent: 'PROTO',
lexicon: [ 'untu', 'unut', 'unət', 'unnu', 'tun', 'əntu' ]
}
]
const tokenizedEpoch = [
{ type: "semicolon", value: "; comment" },
{ type: "star", value: "*" }, { type: "referent", value: "PROTO" }, { type: 'lineBreak', value: '' }, { type: "whiteSpace", value: "" },
{ type: "openBracket", value: "[" }, { type: "plus", value: "+" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "FEATURE" }, { type: "closeBracket", value: "]" },
{ type: "greaterThan", value: ">" }, { type: "openBracket", value: "[" }, { type: "minus", value: "-" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "FEATURE" }, { type: "closeBracket", value: "]" },
{ type: "slash", value: "/" }, { type: "dot", value: "." },
{ type: "underscore", value: "_" }, { type: "dot", value: "." }, { type: 'lineBreak', value: '' }, { type: "whiteSpace", value: "" },
{ type: "referent", value: "n" },
{ type: "greaterThan", value: ">" }, { type: "referent", value: "m" },
{ type: "slash", value: "/" }, { type: "hash", value: "#" },
{ type: "underscore", value: "_" }, { type: "dot", value: "." }, { type: 'lineBreak', value: '' },
{ type: "pipe", value: "|" }, { type: "referent", value: "CHILD" }
]
const treeEpoch = {
epochs: [
{
parent: 'PROTO',
name: 'CHILD',
index: 0,
changes: [
'[+ FEATURE]>[- FEATURE]/._.',
'n>m/#_.'
]
}
]
}
const epochState = {
...initState(),
epochs: treeEpoch.epochs,
latl: epochDefinitionLatl
}
const featureDefinitionLatl = `
[+ PLOSIVE] = kp/p/b/d/t/g/k
[- PLOSIVE] = m/n/s/z
[SONORANT
+= m/n
-= s/z/kp/p/b/d/t/g/k
]
`
const tokenizedFeature = [
{type: "openBracket", value: "[" }, { type: "plus", value: "+" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "PLOSIVE" }, { type: "closeBracket", value: "]" }, { type: "whiteSpace", value: "" },
{ type: "equal", value: "=" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "kp" }, { type: "slash", value: "/" }, { type: "referent", value: "p" }, { type: "slash", value: "/" }, { type: "referent", value: "b" }, { type: "slash", value: "/" }, { type: "referent", value: "d" }, { type: "slash", value: "/" }, { type: "referent", value: "t" }, { type: "slash", value: "/" }, { type: "referent", value: "g" }, { type: "slash", value: "/" }, { type: "referent", value: "k" }, { type: 'lineBreak', value: '' },
{type: "openBracket", value: "[" }, { type: "minus", value: "-" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "PLOSIVE" }, { type: "closeBracket", value: "]" }, { type: "whiteSpace", value: "" },
{ type: "equal", value: "=" }, { type: "whiteSpace", value: "" }, { type: "referent", value: "m" }, { type: "slash", value: "/" }, { type: "referent", value: "n" }, { type: "slash", value: "/" }, { type: "referent", value: "s" }, { type: "slash", value: "/" }, { type: "referent", value: "z" }, { type: 'lineBreak', value: '' },
{type: "openBracket", value: "[" }, { type: "referent", value: "SONORANT" }, { type: 'lineBreak', value: '' },
{ type: "whiteSpace", value: "" }, { type: "positiveAssignment", value: "+=" }, { type: "whiteSpace", value: "" },
{ type: "referent", value: "m" }, { type: "slash", value: "/" }, { type: "referent", value: "n" }, { type: 'lineBreak', value: '' },
{ type: "whiteSpace", value: "" }, { type: "negativeAssignment", value: "-=" }, { type: "whiteSpace", value: "" },
{ type: "referent", value: "s" }, { type: "slash", value: "/" }, { type: "referent", value: "z" }, { type: "slash", value: "/" }, { type: "referent", value: "kp" }, { type: "slash", value: "/" }, { type: "referent", value: "p" }, { type: "slash", value: "/" }, { type: "referent", value: "b" }, { type: "slash", value: "/" }, { type: "referent", value: "d" }, { type: "slash", value: "/" }, { type: "referent", value: "t" }, { type: "slash", value: "/" }, { type: "referent", value: "g" }, { type: "slash", value: "/" }, { type: "referent", value: "k" }, { type: 'lineBreak', value: '' },
{ type: "closeBracket", value: "]" },
]
const treeFeature = { features: [
{
feature: 'PLOSIVE',
positivePhones: ['kp', 'p', 'b', 'd', 't', 'g', 'k'],
negativePhones: ['m', 'n', 's', 'z']
},
{
feature: 'SONORANT',
positivePhones: ['m', 'n'],
negativePhones: ['s' ,'z' ,'kp' ,'p' ,'b' ,'d' ,'t' ,'g' ,'k']
}
]}
const featureState = {
...initState(),
features: {
PLOSIVE: {
negative: [
{
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "m",
},
{
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "n",
},
{
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "s",
},
{
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "z",
},
],
positive: [
{
features: {
PLOSIVE: true,
},
grapheme: "kp",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "p",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "b",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "d",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "t",
ʰ: {
features: {},
grapheme: "tʰ",
},
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "g",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "k",
p: {
features: {
SONORANT: false,
},
grapheme: "kp",
},
},
],
},
SONORANT: {
negative: [
{
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "s",
},
{
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "z",
},
{
features: {
SONORANT: false,
},
grapheme: "kp",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "p",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "b",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "d",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "t",
ʰ: {
features: {},
grapheme: "tʰ",
},
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "g",
},
{
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "k",
p: {
features: {
SONORANT: false,
},
grapheme: "kp",
},
},
],
positive: [
{
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "m",
},
{
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "n",
},
],
}, },
parseResults: 'latl parsed successfully',
latl: featureDefinitionLatl,
phones: {
a: {
features: {},
grapheme: "a",
},
b: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "b",
},
d: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "d",
},
g: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "g",
},
k: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "k",
p: {
features: {
SONORANT: false,
},
grapheme: "kp",
},
},
m: {
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "m",
},
n: {
features: {
PLOSIVE: false,
SONORANT: true,
},
grapheme: "n",
},
p: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "p",
},
s: {
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "s",
},
t: {
features: {
PLOSIVE: true,
SONORANT: false,
},
grapheme: "t",
ʰ: {
features: {},
grapheme: "tʰ",
},
},
u: {
features: {},
grapheme: "u",
},
z: {
features: {
PLOSIVE: false,
SONORANT: false,
},
grapheme: "z",
},
ə: {
features: {},
grapheme: "ə",
},
ɯ: {
features: {},
grapheme: "ɯ",
},
}
}
const lexiconDefinitionLatl = `
/PROTO
kpn
sm
/
`
const tokenizedLexicon = [
{ type: "slash", value: "/" }, { type: "referent", value: "PROTO" }, { type: 'lineBreak', value: '' },
{ type: "whiteSpace", value:"" }, { type: "referent", value: "kpn" }, { type: 'lineBreak', value: '' },
{ type: "whiteSpace", value:"" }, { type: "referent", value: "sm" }, { type: 'lineBreak', value: '' },
{ type: "slash", value: "/" }
]
const treeLexicon = {lexicon: [{epoch: "PROTO", value: ["kpn", "sm"]}]};
const lexiconState = {
...initState(),
latl: lexiconDefinitionLatl,
lexicon: [
{ lexeme: 'kpn', epoch: 'PROTO'},
{ lexeme: 'sm', epoch: 'PROTO'}
],
parseResults: 'latl parsed successfully'
}
const totalLatl = `${epochDefinitionLatl}\n\n${featureDefinitionLatl}\n\n${lexiconDefinitionLatl}`
const totalLatlState = {
...initState(),
latl: totalLatl,
phonemes: {},
features: featureState.features,
epochs: treeEpoch.epochs,
lexicon: lexiconState.lexicon,
parseResults: 'latl parsed successfully'
}

View file

@ -1,293 +0,0 @@
// @flow
import type { stateType, epochType, phoneType } from './reducer';
export type resultsAction = {
type: 'RUN'
}
export type decomposedRulesType = [
{
environment: {
pre: [{[key: string]: boolean}],
position: [{[key: string]: boolean}],
post: [{[key: string]: boolean}]
},
newFeatures: [{[key: string]: boolean}]
}
]
type ruleBundle = {
environment: {
pre: string,
position: string,
post: string
},
newFeatures: string
}
const getProperty = property => object => object[property]
const findFeaturesFromLexeme = (phones: {}, lexeme:string): [] => {
let featureBundle = []
let lastIndex = lexeme.length - 1;
let node = {};
[...lexeme].forEach((graph, index) => {
try {
if (!index ) return node = phones[graph]
if (index === lastIndex) return node[graph]
? featureBundle.push(node[graph])
: featureBundle.push(node, phones[graph])
if (!node[graph] && node.features) {
featureBundle.push(node)
return node = phones[graph]
}
if (!node) return node = phones[graph]
return node = node[graph]
}
catch (e) {
throw {e, 'phones[graph]':phones[graph], index, lexeme }
}
})
return featureBundle;
}
const findFeaturesFromGrapheme = (phones: {}, lexeme:string): [] => {
let featureBundle = []
let lastIndex = lexeme.length - 1;
let node = {};
[...lexeme].forEach((graph, index) => {
if (!index && !lastIndex) featureBundle.push(phones[graph].features)
if (!index) return node = phones[graph]
if (index === lastIndex) return node[graph]
? featureBundle.push(node[graph])
: featureBundle.push(node, phones[graph])
if (!node[graph] && node.features) {
featureBundle.push(node)
return node = phones[graph]
}
if (!node[graph])
return node = node[graph]
})
return featureBundle;
}
const errorMessage = ([prefix, separator], location, err) => `${prefix}${location}${separator}${err}`
const lintRule = (rule) => {
if (!rule.match(/>/g)) throw `Insert '>' operator between target and result`
if (!rule.match(/\//g)) throw `Insert '/' operator between change and environment`
if (!rule.match(/_/g)) throw `Insert '_' operator in environment`
if (rule.match(/>/g).length > 1) throw `Too many '>' operators`
if (rule.match(/\//g).length > 1) throw `Too many '/' operators`
if (rule.match(/_/g).length > 1) throw `Too many '_' operators`
return rule.split(/>|\/|_/g);
}
const decomposeRule = (rule: string, index: number): ruleBundle => {
try {
// splits rule at '>' '/' and '_' substrings resulting in array of length 4
const [position, newFeatures, pre, post] = lintRule(rule);
return { environment: { pre, position, post }, newFeatures }
} catch (err) {
throw errorMessage`Error in line ${index + 1}: ${err}`;
}
}
const isUnknownFeatureToken = token => token !== '-' && token !== '+' && token !== ']' && token !== '[' && token !== ' ';
const doesFeatureRuleContainUnknownToken = features => {
const unknownTokens = features
.match(/\W/g)
.filter(isUnknownFeatureToken)
if (unknownTokens.length) throw `Unknown token '${unknownTokens[0]}'`;
return true
}
const reduceFeaturesToBoolean = bool => (map, feature) => ({...map, [feature]: bool})
const getFeatures = (phoneme: string, featureBoolean): {} => {
try {
const featureMatch = featureBoolean
// regEx to pull positive features
? /(?=\+.).*(?<=\-)|(?=\+.).*(?!\-).*(?<=\])/g
// regEx to pull negative features
: /(?=\-.).*(?<=\+)|(?=\-.).*(?!\+).*(?<=\])/g
const [ features ] = phoneme.match(featureMatch) || [ null ];
if (features) {
doesFeatureRuleContainUnknownToken(features)
return features
.trim()
.match(/\w+/g)
.reduce(reduceFeaturesToBoolean(featureBoolean), {})
}
return {}
} catch (err) {
throw err;
}
}
const mapToPositiveAndNegativeFeatures = phoneme => (
{ ...getFeatures(phoneme, true), ...getFeatures(phoneme, false) } )
const mapStringToFeatures = (ruleString, phones) => {
if (ruleString) {
if (ruleString === '.') return [];
if (ruleString === '#') return ['#']
if (ruleString === '0') return [];
const ruleBrackets = ruleString.match(/\[.*\]/)
try {
if (ruleBrackets) {
return ruleString
.split('[')
// filter out empty strings
.filter(v => v)
.map(mapToPositiveAndNegativeFeatures)
}
return findFeaturesFromGrapheme(phones, ruleString);
} catch (err) {
throw err;
}
}
return {};
}
const mapRuleBundleToFeatureBundle = phones => ( ruleBundle, index ) => {
// for each object in ruleBundle, map values to array of objects with feature-boolean key-value pairs
try {
const { newFeatures, environment:{ pre, position, post } } = ruleBundle;
return {
environment: {
pre: mapStringToFeatures(pre, phones),
position: mapStringToFeatures(position, phones),
post: mapStringToFeatures(post, phones),
},
newFeatures: mapStringToFeatures(newFeatures, phones)
}
} catch (err) {
throw errorMessage`Error in line ${index + 1}: ${err}`;
}
}
export const decomposeRules = (epoch: epochType, phones: {[key: string]: phoneType}): decomposedRulesType => {
const { changes } = epoch
try {
return changes
.map(decomposeRule)
.map(mapRuleBundleToFeatureBundle(phones));
} catch (err) {
const ruleError = {epoch: epoch.name, error: err}
throw ruleError;
}
}
const isPhonemeBoundByRule = phonemeFeatures => (ruleFeature, index) => {
const phoneme = phonemeFeatures[index].features;
return Object.entries(ruleFeature).reduce((bool, [feature, value]) => {
if (!bool) return false;
if (!phoneme.hasOwnProperty(feature)) return false;
if (!phoneme[feature] && !value) return true;
if (phoneme[feature] !== value) return false;
return true;
}, true);
}
const isEnvironmentBoundByRule = (phonemeFeatures, ruleFeatures) => {
if (!ruleFeatures) return true;
return ruleFeatures.filter(isPhonemeBoundByRule(phonemeFeatures)).length === ruleFeatures.length;
}
const getEntries = object => Object.entries(object);
const isObjectWithPropertyInArray = (array, property) => candidate => array.map(getProperty(property)).includes(candidate[property]);
const transformFeatureValues = features => ([newFeature, newValue]) => features[newFeature][newValue ? 'positive': 'negative'];
const reduceFeatureValues = (newPhoneme, [newFeature, newValue]) => ({ ...newPhoneme, [newFeature]: newValue })
const transformPhoneme = (phoneme, newFeatures, features) => {
if (!newFeatures) return {}
const newPhonemeFeatures = getEntries(newFeatures).reduce(reduceFeatureValues, {...phoneme.features});
const newPhonemeCandidates = getEntries(newPhonemeFeatures).map(transformFeatureValues(features));
return newPhonemeCandidates
.reduce((candidates, candidatesSubset, index, array) => candidates.filter(isObjectWithPropertyInArray(candidatesSubset, 'grapheme'))
, newPhonemeCandidates[newPhonemeCandidates.length - 1])[0];
}
const transformLexemeInitial = (newLexeme, pre, post, position, phoneme, index, lexemeBundle, newFeatures, features) => {
if (index !== pre.length - 1) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule([phoneme], position)) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule(lexemeBundle.slice(index + position.length, index + post.length + position.length), post)) return [...newLexeme, phoneme];
const newPhoneme = transformPhoneme(phoneme, newFeatures[0], features);
// if deletion occurs
if (!newPhoneme || !newPhoneme.grapheme) return [ ...newLexeme] ;
return [...newLexeme, newPhoneme];
}
const transformLexemeCoda = (newLexeme, pre, post, position, phoneme, index, lexemeBundle, newFeatures, features) => {
if (index + post.length !== lexemeBundle.length) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule(lexemeBundle.slice(index - pre.length, index), pre)) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule([phoneme], position)) return [...newLexeme, phoneme];
const newPhoneme = transformPhoneme(phoneme, newFeatures[0], features);
// if deletion occurs
if (!newPhoneme || !newPhoneme.grapheme) return [ ...newLexeme] ;
return [...newLexeme, newPhoneme];
}
export const transformLexeme = (lexemeBundle, rule, features) => {
const {pre, post, position} = rule.environment;
const newLexeme = lexemeBundle.reduce((newLexeme, phoneme, index) => {
if (pre.find(val => val === '#')) return transformLexemeInitial(newLexeme, pre, post, position, phoneme, index, lexemeBundle, rule.newFeatures, features);
if (post.find(val => val === '#')) return transformLexemeCoda(newLexeme, pre, post, position, phoneme, index, lexemeBundle, rule.newFeatures, features);
if ( index < pre.length || index >= lexemeBundle.length - post.length ) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule(lexemeBundle.slice(index - pre.length, index), pre)) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule([phoneme], position)) return [...newLexeme, phoneme];
if (!isEnvironmentBoundByRule(lexemeBundle.slice(index, index + post.length), post)) return [...newLexeme, phoneme];
const newPhoneme = transformPhoneme(phoneme, rule.newFeatures[0], features);
// if deletion occurs
if (!newPhoneme || !newPhoneme.grapheme) return [ ...newLexeme] ;
return [...newLexeme, newPhoneme];
}, [])
return newLexeme;
}
const formBundleFromLexicon = lexicon => phones => lexicon.map(({lexeme}) => findFeaturesFromLexeme(phones, lexeme))
const transformLexicon = lexiconBundle =>
ruleBundle =>
features =>
lexiconBundle.map(lexemeBundle => ruleBundle.reduce(
(lexeme, rule, i) => transformLexeme(lexeme, rule, features)
, lexemeBundle
))
const getGraphemeFromEntry = ([_, phoneme]) => phoneme.grapheme
const stringifyLexeme = (lexeme) => lexeme.map(getProperty('grapheme')).join('')
const stringifyResults = ({lexicon, ...passResults}) => ({...passResults, lexicon: lexicon.map(stringifyLexeme)})
export const run = (state: stateType, action: resultsAction): stateType => {
// TODO iterate through each epoch
try {
const passResults = state.epochs.reduce((results, epoch, _) => {
const { phones, features, lexicon } = state;
let lexiconBundle;
if ( epoch.parent ) {
lexiconBundle = results.find(result => result.pass === epoch.parent)
}
if (!lexiconBundle) {
lexiconBundle = formBundleFromLexicon(lexicon)(phones);
}
else {
lexiconBundle = lexiconBundle.lexicon
}
const ruleBundle = decomposeRules(epoch, phones);
const passResults = transformLexicon(lexiconBundle)(ruleBundle)(features)
const pass = { pass: epoch.name, lexicon: passResults }
if ( epoch.parent ) pass.parent = epoch.parent;
return [...results, pass];
}, []);
const results = passResults.map(stringifyResults);
return {...state, results, errors: {}, parseResults: '' }
} catch (err) {
console.log(err)
return {...state, errors: err, results:[], parseResults: '' };
}
}

View file

@ -1,318 +0,0 @@
import { stateReducer } from './reducer';
import { initState } from './reducer.init';
import { decomposeRules, transformLexeme } from './reducer.results';
describe('Results', () => {
let state = {};
beforeEach(()=> {
state = {};
})
it('results returned unaltered', () => {
const action = {type: ''};
expect(stateReducer(state, action)).toBe(state);
});
it('rules decomposed properly', () => {
const { epochs, phones } = initState(1);
const result = getResult();
expect(decomposeRules(epochs[0], phones)).toStrictEqual(result);
});
it('rule without ">" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't/n/_' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Insert '>' operator between target and result"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule with too many ">" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't>n>/_' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Too many '>' operators"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule without "/" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't>n_' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Insert '/' operator between change and environment"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule with too many "/" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't>n/_/' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Too many '/' operators"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule without "_" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't>n/' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Insert '_' operator in environment"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule with too many "_" returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ 't>n/__' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Too many '_' operators"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('rule with incorrect feature syntax returns helpful error message', () => {
const { phones } = initState();
const epoch = { name: 'error epoch', changes: [ '[+ occlusive - nasal = obstruent]>n/_' ] }
const errorMessage = {epoch: 'error epoch', error: "Error in line 1: Unknown token '='"};
let receivedError;
try {
decomposeRules(epoch, phones)
}
catch (err) {
receivedError=err;
}
expect(receivedError).toStrictEqual(errorMessage);
})
it('expect transform lexeme to apply rule to lexeme', () => {
const lexemeBundle = getlexemeBundle();
const resultsLexeme = [...lexemeBundle]
resultsLexeme[2] = lexemeBundle[1]
const rule = getRule();
expect(transformLexeme(lexemeBundle, rule, initState().features)).toEqual(resultsLexeme)
})
it('results returned from first sound change rule (feature matching)', () => {
const action = {type: 'RUN'};
state = initState(1)
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'anna', 'anat', 'anət', 'anna', 'tan', 'ənna'
]
}
]);
});
it('results returned through second sound change rule (phoneme matching)', () => {
const action = {type: 'RUN'};
state = initState(2)
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annɯ', 'anat', 'anət', 'annɯ', 'tan', 'ənnɯ'
]
}
]);
});
it('results returned through third sound change rule (phoneme dropping)', () => {
const action = {type: 'RUN'};
state = initState(3)
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annɯ', 'anat', 'ant', 'annɯ', 'tan', 'nnɯ'
]
}
]);
});
it('results returned through fourth sound change rule (lexeme initial environment)', () => {
const action = {type: 'RUN'};
state = initState(4)
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annɯ', 'anat', 'ant', 'annɯ', 'tʰan', 'nnɯ'
]
}
]);
});
it('results returned through fifth sound change rule (lexeme final environment)', () => {
const action = {type: 'RUN'};
state = initState(5)
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annu', 'anat', 'ant', 'annu', 'tʰan', 'nnu'
]
}
]);
});
// it('results returned through sixth sound change rule (multi-phoneme target)', () => {
// const action = {type: 'RUN'};
// state = initState(6)
// expect(stateReducer(state, action).results).toEqual([
// {
// pass: 'epoch-1',
// lexicon: [
// 'annu', 'anta', 'ant', 'annu', 'tʰan', 'nnu'
// ]
// }
// ]);
// });
it('results returned for multiple epochs without parent epoch', () => {
const action = {type: 'RUN'};
state = initState(5);
const newEpoch = {
name: 'epoch-2',
changes: [
'[+ sonorant ]>0/#_.',
'n>0/#_n'
]
}
state.epochs = [ ...state.epochs, newEpoch ]
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annu', 'anat', 'ant', 'annu', 'tʰan', 'nnu'
]
},
{
pass: 'epoch-2',
lexicon: [
'nta', 'nat', 'nət', 'na', 'tan', 'nta'
]
}
])
})
it('results returned for multiple epochs with parent epoch', () => {
const action = {type: 'RUN'};
state = initState(5);
const newEpoch = {
name: 'epoch-2',
parent: 'epoch-1',
changes: [
'[+ sonorant ]>0/#_.'
]
}
state.epochs = [ ...state.epochs, newEpoch ]
expect(stateReducer(state, action).results).toEqual([
{
pass: 'epoch-1',
lexicon: [
'annu', 'anat', 'ant', 'annu', 'tʰan', 'nnu'
]
},
{
pass: 'epoch-2',
parent: 'epoch-1',
lexicon: [
'nnu', 'nat', 'nt', 'nnu', 'tʰan', 'nu'
]
}
])
})
});
const getlexemeBundle = () => ([
{
grapheme: 'a',
features: {
sonorant: true,
back: true,
low: true,
high: false,
rounded: false
}
},
{
grapheme: 'n',
features: { sonorant: true, nasal: true, occlusive: true, coronal: true }
},
{
grapheme: 't',
features: { occlusive: true, coronal: true, obstruent: true, nasal: false }
},
{
grapheme: 'a',
features: {
sonorant: true,
back: true,
low: true,
high: false,
rounded: false
}
}
])
const getRule = () => ({
environment: {
pre: [ { sonorant: true, nasal: true, occlusive: true, coronal: true } ],
position: [ { occlusive: true, nasal: false } ],
post: []
},
newFeatures: [ { occlusive: true, nasal: true } ]
})
const getResult = () => ([
{
environment: {
pre: [
{
sonorant: true, nasal: true, occlusive: true, coronal: true
}
],
position: [
{occlusive: true, nasal: false}
],
post: [],
},
newFeatures: [{occlusive: true, nasal: true}]
}
]);

View file

@ -1,18 +1,17 @@
// @flow // @flow
import type { stateType } from './reducer'; import type { stateType } from './stateReducer';
export type epochAction = { export type epochAction = {
type: "ADD_EPOCH" | "SET_EPOCH" | "REMOVE_EPOCH", type: "ADD_EPOCH" | "SET_EPOCH" | "REMOVE_EPOCH",
value: { value: {
index?: number, index?: number,
name: string, name: string,
changes?: Array<string>, changes?: Array<string>
parent?: string
} }
} }
export const addEpoch = (state: stateType, action: epochAction): stateType => { export const addEpoch = (state: stateType, action: epochAction): stateType => {
const newEpoch = { name: action.value.name, changes: action.value.changes || [''], parent: null}; const newEpoch = { name: action.value.name, changes: action.value.changes || [''] };
return {...state, epochs: [...state.epochs, newEpoch]} return {...state, epochs: [...state.epochs, newEpoch]}
} }
@ -28,10 +27,6 @@ export const setEpoch = (state: stateType, action: epochAction): stateType => {
mutatedEpochs[index].changes = action.value.changes mutatedEpochs[index].changes = action.value.changes
? action.value.changes ? action.value.changes
: mutatedEpochs[index].changes; : mutatedEpochs[index].changes;
mutatedEpochs[index].parent = action.value.parent && action.value.parent !== 'none'
? action.value.parent
: null
return {...state, epochs: [...mutatedEpochs]} return {...state, epochs: [...mutatedEpochs]}
} }

View file

@ -1,13 +1,12 @@
import {stateReducer} from './reducer'; import {stateReducer} from './stateReducer';
describe('Epochs', () => { describe('Epochs', () => {
const state = {}; const state = {};
beforeEach(()=> { beforeEach(()=> {
state.epochs = [ state.epochs = [
{ {
name: 'epoch-1', name: 'epoch 1',
changes: [''], changes: ['']
parent: null
} }
] ]
}) })
@ -18,45 +17,45 @@ describe('Epochs', () => {
}); });
it('epochs addition returns new epochs list', () => { it('epochs addition returns new epochs list', () => {
const action = {type: 'ADD_EPOCH', value: { name: 'epoch-2', changes: [''], parent: null}}; const action = {type: 'ADD_EPOCH', value: { name: 'epoch 2', changes: ['']}};
expect(stateReducer(state, action)).toEqual({...state, epochs: [...state.epochs, action.value]}) expect(stateReducer(state, action)).toEqual({...state, epochs: [...state.epochs, action.value]})
}) })
it('epoch-name mutation returns new epochs list with mutation', () => { it('epoch name mutation returns new epochs list with mutation', () => {
const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch-2', changes: ['']}}; const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch 2', changes: ['']}};
const secondAction = {type: 'SET_EPOCH', value: { index: 0, name: 'proto-lang'}}; const secondAction = {type: 'SET_EPOCH', value: { index: 0, name: 'proto-lang'}};
const secondState = stateReducer(state, firstAction); const secondState = stateReducer(state, firstAction);
expect(stateReducer(secondState, secondAction)).toEqual( expect(stateReducer(secondState, secondAction)).toEqual(
{...state, {...state,
epochs: [ epochs: [
{name: 'proto-lang', changes: [''], parent: null}, {name: 'proto-lang', changes: ['']},
{name: 'epoch-2', changes: [''], parent: null} {name: 'epoch 2', changes: ['']}
] ]
} }
); );
}); });
it('epoch changes mutation returns new epochs list with mutation', () => { it('epoch changes mutation returns new epochs list with mutation', () => {
const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch-2', changes: ['']}}; const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch 2', changes: ['']}};
const secondAction = {type: 'SET_EPOCH', value: { index: 0, changes: ['n>t/_#', '[+plosive]>[+nasal -plosive]/_n']}}; const secondAction = {type: 'SET_EPOCH', value: { index: 0, changes: ['n>t/_#', '[+plosive]>[+nasal -plosive]/_n']}};
const secondState = stateReducer(state, firstAction); const secondState = stateReducer(state, firstAction);
expect(stateReducer(secondState, secondAction)).toEqual( expect(stateReducer(secondState, secondAction)).toEqual(
{...state, {...state,
epochs: [ epochs: [
{name: 'epoch-1', changes: ['n>t/_#', '[+plosive]>[+nasal -plosive]/_n'], parent: null}, {name: 'epoch 1', changes: ['n>t/_#', '[+plosive]>[+nasal -plosive]/_n']},
{name: 'epoch-2', changes: [''], parent: null} {name: 'epoch 2', changes: ['']}
] ]
} }
); );
}); });
it('epochs returned with deleted epoch removed', () => { it('epochs returned with deleted epoch removed', () => {
const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch-2', changes: ['']}}; const firstAction = {type: 'ADD_EPOCH', value: { name: 'epoch 2', changes: ['']}};
const stateWithTwoEpochs = stateReducer(state, firstAction); const stateWithTwoEpochs = stateReducer(state, firstAction);
const secondAction = {type: 'REMOVE_EPOCH', value: {index: 0, name: 'epoch-1'}} const secondAction = {type: 'REMOVE_EPOCH', value: {index: 0, name: 'epoch 1'}}
expect(stateReducer(stateWithTwoEpochs, secondAction)).toEqual({ expect(stateReducer(stateWithTwoEpochs, secondAction)).toEqual({
...state, ...state,
epochs: [{ name: 'epoch-2', changes: [''], parent: null}] epochs: [{ name: 'epoch 2', changes: ['']}]
}); });
}); });

View file

@ -1,5 +1,5 @@
// @flow // @flow
import type { stateType } from './reducer'; import type { stateType } from './stateReducer';
export type featureAction = { export type featureAction = {
type: "ADD_FEATURE", type: "ADD_FEATURE",
@ -12,45 +12,33 @@ export type featureAction = {
const addPhones = (phones: {}, phone: string): {} => { const addPhones = (phones: {}, phone: string): {} => {
let node = {}; let node = {};
phone.split('').forEach((graph, index) => { phone.split('').forEach((graph, index) => {
if (index) node[graph] = {} if (index) node[graph] = {}
if (!index && !phones[graph]) phones[graph] = {} if (!index && !phones[graph]) phones[graph] = {}
node = index === 0 ? phones[graph] : node[graph]; node = index === 0 ? phones[graph] : node[graph];
if (index === phone.length - 1) node.grapheme = phone; if (index === phone.length - 1) node.grapheme = phone;
}) })
return phones; return phones;
} }
const findPhone = (phones: {}, phone: string): {} => { const findPhone = (phones: {}, phone: string): {} => {
return phone let node = {};
.split('') phone.split('').forEach((graph, index) => {
.reduce((node, graph, index) => {
node = index === 0 ? phones[graph] : node[graph]; node = index === 0 ? phones[graph] : node[graph];
});
return node; return node;
}, {});
} }
const addFeatureToPhone = ( const addFeatureToPhone = (
phones: {}, phone: string, featureKey: string, featureValue: boolean phones: {}, phone: string, featureKey: string, featureValue: boolean
): {} => { ): {} =>
try { {
let node = {} let node = {}
phone.split('').forEach((graph, index) => { phone.split('').forEach((graph, index) => {
node = index === 0 ? phones[graph] : node[graph]; node = index === 0 ? phones[graph] : node[graph];
if (index === phone.split('').length - 1) node.features = {...node.features, [featureKey]: featureValue}
if (index === phone.split('').length - 1) { })
node.features = node && node.features
? {...node.features, [featureKey]: featureValue }
: {[featureKey]: featureValue};
}
});
return phones; return phones;
}
catch (e) {
throw { phones, phone, featureKey, featureValue }
}
} }
export const addFeature = (state: stateType, action: featureAction): stateType => { export const addFeature = (state: stateType, action: featureAction): stateType => {
@ -59,8 +47,7 @@ export const addFeature = (state: stateType, action: featureAction): stateType =
let newFeatureName = action.value.feature; let newFeatureName = action.value.feature;
let newPhoneObject = [ let newPhoneObject = [
...positivePhones, ...negativePhones ...positivePhones, ...negativePhones
] ].reduce((phoneObject, phone) => addPhones(phoneObject, phone), state.phones)
.reduce((phoneObject, phone) => addPhones(phoneObject, phone), state.phones)
if (positivePhones) { if (positivePhones) {
@ -85,11 +72,3 @@ export const addFeature = (state: stateType, action: featureAction): stateType =
let newFeature = {[action.value.feature]: {positive: positivePhones, negative: negativePhones}}; let newFeature = {[action.value.feature]: {positive: positivePhones, negative: negativePhones}};
return {...state, features:{...state.features, ...newFeature}, phones: newPhoneObject} return {...state, features:{...state.features, ...newFeature}, phones: newPhoneObject}
} }
export const deleteFeature = (state, action) => {
const deletedFeature = state.features[action.value];
deletedFeature.positive.forEach(phone => delete phone.features[action.value])
deletedFeature.negative.forEach(phone => delete phone.features[action.value])
delete state.features[action.value];
return state
}

View file

@ -1,4 +1,4 @@
import {stateReducer} from './reducer'; import {stateReducer} from './stateReducer';
describe('Features', () => { describe('Features', () => {
const state = {} const state = {}
@ -31,17 +31,4 @@ describe('Features', () => {
); );
}); });
it('feature deletion returns new feature list', () => {
const action = {type: 'DELETE_FEATURE', value: 'occlusive'}
expect(stateReducer(state, action)).toEqual(
{...state,
features: {},
phones: {
a: {features: {}, grapheme: 'a'},
n: {features: {}, grapheme: 'n'}
}
}
)
})
}); });

View file

@ -0,0 +1,92 @@
// @flow
import type { stateType } from './stateReducer';
export type initAction = {
type: "INIT"
}
export const initState = (changesArgument: number = -1): stateType => {
const state = {
epochs: [
{
name: 'epoch 1',
changes: [
'[+ occlusive - nasal]>[+ occlusive nasal]/n_',
'at>ta/_#',
'[+ sonorant - low rounded high back]>_/_',
'nn>nun/_',
'[+ nasal][+ obstruent]>[+ nasal obstruent aspirated ]/#_',
'[+ sonorant rounded]>[+ sonorant - rounded]/_#'
]
}
],
phones: {
a: {
grapheme: 'a', features: {
sonorant: true, back: true, low: true, high: false, rounded: false
}
},
u: {
grapheme: 'u', features: {
sonorant: true, back: true, low: false, high: true, rounded: true,
}
},
ɯ: {
grapheme: 'ɯ', features: {
sonorant: true, back: true, low: false, high: true, rounded: false,
}
},
ə: {
grapheme: 'ə', features: {
sonorant: true, low: false, rounded: false, high: false, back: false
}
},
t: {
grapheme: 't', features: {
occlusive: true, coronal: true, obstruent: true
},
ʰ: {
grapheme: 'tʰ', features: {
occlusive: true, coronal: true, obstruent: true, aspirated: true
}
}
},
n: {
grapheme: 'n', features: {
sonorant: true, nasal: true, occlusive: true, coronal: true
}
}
},
options: {
output: 'default', save: false
},
results: {},
errors: {},
features: {},
lexicon: []
};
state.features = {
sonorant: { positive:[ state.phones.a, state.phones.u, state.phones.ɯ, state.phones.ə, state.phones.n], negative: [] },
back: { positive:[ state.phones.a, state.phones.u, state.phones.ɯ ], negative: [ state.phones.ə ] },
low: { positive:[ state.phones.a ], negative: [ state.phones.u, state.phones.ɯ, state.phones.ə ] },
high: { positive:[ state.phones.u, state.phones.ɯ ], negative: [ state.phones.a, state.phones.ə ] },
rounded: { positive:[ state.phones.u ], negative: [ state.phones.a, state.phones.ɯ, state.phones.ə ] },
occlusive: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
coronal: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
obstruent: { positive:[ state.phones.t, state.phones.n, state.phones.t.ʰ ], negative: [] },
nasal: { positive:[ state.phones.n ], negative: [] },
aspirated: { positive:[ state.phones.t.ʰ ], negative: [] },
}
state.lexicon = [
{lexeme: 'anta', epoch: state.epochs[0]},
{lexeme: 'anat', epoch: state.epochs[0]},
{lexeme: 'anət', epoch: state.epochs[0]},
{lexeme: 'anna', epoch: state.epochs[0]},
{lexeme: 'tan', epoch: state.epochs[0]},
{lexeme: 'ənta', epoch: state.epochs[0]}
]
if(changesArgument > -1) state.epochs[0].changes = state.epochs[0].changes.splice(changesArgument, 1)
return state;
}

View file

@ -1,25 +1,23 @@
// @flow // @flow
import { addLexeme, setLexicon } from './reducer.lexicon'; import { addLexeme, setLexicon } from './stateReducer.lexicon';
import type { lexiconAction } from './reducer.lexicon'; import type { lexiconAction } from './stateReducer.lexicon';
import { addEpoch, setEpoch, removeEpoch } from './reducer.epochs'; import { addEpoch, setEpoch, removeEpoch } from './stateReducer.epochs';
import type { epochAction } from './reducer.epochs'; import type { epochAction } from './stateReducer.epochs';
import { addFeature, deleteFeature } from './reducer.features'; import { addFeature } from './stateReducer.features';
import type { featureAction } from './reducer.features'; import type { featureAction } from './stateReducer.features';
import type { optionsAction } from './reducer.options'; import type { optionsAction } from './stateReducer.options';
import { setOptions } from './reducer.options'; import { setOptions } from './stateReducer.options';
import { run } from './reducer.results'; import { run } from './stateReducer.results';
import type { resultsAction } from './reducer.results' import type { resultsAction } from './stateReducer.results'
import { initState } from './reducer.init'; import { initState } from './stateReducer.init';
import type { initAction } from './reducer.init'; import type { initAction } from './stateReducer.init';
import { clearOutput } from './reducer.clear';
import { setLatl, parseLatl } from './reducer.latl';
export type stateType = { export type stateType = {
lexicon: Array<{lexeme: string, epoch: epochType}>, lexicon: Array<{lexeme: string, epoch: epochType}>,
epochs: Array<epochType>, epochs: Array<epochType>,
phones: {[key: string]: phoneType}, phones: {[key: string]: phoneType},
options: {output: string, save: boolean}, options: {output: string, save: boolean},
results: [], results: {},
errors: {}, errors: {},
features: featureType features: featureType
} }
@ -57,16 +55,8 @@ export const stateReducer = (state: stateType, action: actionType): stateType =>
case 'ADD_FEATURE': return addFeature(state, action); case 'ADD_FEATURE': return addFeature(state, action);
case 'DELETE_FEATURE': return deleteFeature(state, action);
case 'SET_OPTIONS': return setOptions(state, action); case 'SET_OPTIONS': return setOptions(state, action);
case 'SET_LATL': return setLatl(state, action);
case 'PARSE_LATL': return parseLatl(state, action);
case 'CLEAR': return clearOutput(state, action);
case 'RUN': return run(state, action); case 'RUN': return run(state, action);
default: return state; default: return state;

View file

@ -1,5 +1,5 @@
// @flow // @flow
import type { stateType } from './reducer'; import type { stateType } from './stateReducer';
type lexemeType = { type lexemeType = {
lexeme: string, lexeme: string,
@ -20,10 +20,8 @@ const makeLexeme = (lexeme: string, epochName: ?string, state: stateType) => {
const newLexeme = {lexeme: lexeme, epoch: state.epochs[0]}; const newLexeme = {lexeme: lexeme, epoch: state.epochs[0]};
if (epochName) { if (epochName) {
const epochIndex = state.epochs.findIndex(epoch => epoch.name === epochName); const epochIndex = state.epochs.findIndex(epoch => epoch.name === epochName);
if (epochIndex > -1) { if (epochIndex > 0) {
newLexeme.epoch = state.epochs[epochIndex]; newLexeme.epoch = state.epochs[epochIndex];
} else {
newLexeme.epoch = epochName;
}; };
} }
return newLexeme; return newLexeme;

View file

@ -1,10 +1,10 @@
import {stateReducer} from './reducer'; import {stateReducer} from './stateReducer';
describe('Lexicon', () => { describe('Lexicon', () => {
const state = { const state = {
epochs: [ epochs: [
{ name: 'epoch-1', changes:[''] }, { name: 'epoch 1', changes:[''] },
{ name: 'epoch-2', changes:[''] } { name: 'epoch 2', changes:[''] }
] ]
} }
state.lexicon = [ state.lexicon = [
@ -28,16 +28,16 @@ describe('Lexicon', () => {
}); });
it('lexicon addition with epoch returns updated lexicon with correct epoch', () => { it('lexicon addition with epoch returns updated lexicon with correct epoch', () => {
const action = {type: 'ADD_LEXEME', value: {lexeme:'ntʰa', epoch: 'epoch-2'}} const action = {type: 'ADD_LEXEME', value: {lexeme:'ntʰa', epoch: 'epoch 2'}}
expect(stateReducer(state, action)).toEqual({...state, lexicon:[...state.lexicon, {lexeme:'ntʰa', epoch:state.epochs[1]}]}); expect(stateReducer(state, action)).toEqual({...state, lexicon:[...state.lexicon, {lexeme:'ntʰa', epoch:state.epochs[1]}]});
}); });
it('lexicon set returns updated lexicon with correct epoch', () => { it('lexicon set returns updated lexicon with correct epoch', () => {
const newLexicon = [ const newLexicon = [
{lexeme:'anta', epoch:'epoch-1'}, {lexeme:'anta', epoch:'epoch 1'},
{lexeme:'anat', epoch:'epoch-1'}, {lexeme:'anat', epoch:'epoch 1'},
{lexeme:'anət', epoch:'epoch-1'}, {lexeme:'anət', epoch:'epoch 1'},
{lexeme:'anna', epoch:'epoch-1'} {lexeme:'anna', epoch:'epoch 1'}
] ]
const action = {type: 'SET_LEXICON', value: newLexicon} const action = {type: 'SET_LEXICON', value: newLexicon}
expect(stateReducer(state, action)).toEqual({...state, lexicon:[ expect(stateReducer(state, action)).toEqual({...state, lexicon:[
@ -58,7 +58,7 @@ describe('Lexicon', () => {
const inputLexicon = [ const inputLexicon = [
{lexeme:'anta'}, {lexeme:'anta'},
{lexeme:'anat'}, {lexeme:'anat'},
{lexeme:'anət', epoch:'epoch-2'}, {lexeme:'anət', epoch:'epoch 2'},
{lexeme:'anna'} {lexeme:'anna'}
] ]
const action = {type: 'SET_LEXICON', value: inputLexicon} const action = {type: 'SET_LEXICON', value: inputLexicon}

View file

@ -1,5 +1,5 @@
// @flow // @flow
import type { stateType } from './reducer'; import type { stateType } from './stateReducer';
export type optionAction = { export type optionAction = {
type: 'SET_OPTIONS', type: 'SET_OPTIONS',

View file

@ -1,5 +1,5 @@
import { stateReducer } from './reducer'; import { stateReducer } from './stateReducer';
import { initState } from './reducer.init'; import { initState } from './stateReducer.init';
describe('Options', () => { describe('Options', () => {
let state = {} let state = {}

View file

@ -1,4 +1,4 @@
import {stateReducer} from './reducer'; import {stateReducer} from './stateReducer';
describe('Phones', () => { describe('Phones', () => {
const n_phone = {features: {nasal: true}, grapheme: 'n'}; const n_phone = {features: {nasal: true}, grapheme: 'n'};

View file

@ -0,0 +1,57 @@
// @flow
import type { stateType } from './stateReducer';
export type resultsAction = {
type: 'RUN'
}
const findFeatures = (phones: {}, lexeme:string): [] => {
let featureBundle = []
let lastIndex = lexeme.length - 1;
let node = {};
[...lexeme].forEach((graph, index) => {
if (!index) return node = phones[graph]
if (index === lastIndex) return node[graph]
? featureBundle.push(node[graph])
: featureBundle.push(node, phones[graph])
if (!node[graph] && node.features) {
featureBundle.push(node)
return node = phones[graph]
}
if (!node[graph])
return node = node[graph]
})
return featureBundle;
}
const decomposeRule = (rule: string): string[] => {
let decomposedChange = rule.split('>');
decomposedChange = [decomposedChange[0], ...decomposedChange[1].split('/')]
decomposedChange = [decomposedChange[0], decomposedChange[1], ...decomposedChange[2].split('_')];
return [...decomposedChange];
}
export const run = (state: stateType, action: resultsAction): stateType => {
// ! one epoch only
// rule 0 '[+ occlusive - nasal]>[+ occlusive nasal]/n_'
let ruleBundle = state.epochs[0].changes;
ruleBundle = ruleBundle.map(rule => decomposeRule(rule))
ruleBundle.map(rule => {
rule.forEach(position => {
console.log(position)
})
})
let featurePhoneBundle = state.lexicon.map(lexeme => findFeatures(state.phones, lexeme))
console.log(featurePhoneBundle)
ruleBundle.forEach(rule => {
featurePhoneBundle.map(featurePhone => {
// if (findRules(featurePhone, )
})
})
let results = [];
return {...state, results: { pass: state.epochs[0].name, results } }
}

View file

@ -0,0 +1,26 @@
import {stateReducer} from './stateReducer';
import {initState} from './stateReducer.init';
describe('Results', () => {
let state = {};
beforeEach(()=> {
state = {};
})
it('results returned unaltered', () => {
const action = {type: ''};
expect(stateReducer(state, action)).toBe(state);
});
it('results returned from first sound change rule', () => {
const action = {type: 'RUN'};
state = initState(0)
expect(stateReducer(state, action).results).toEqual({
pass: 'epoch 1',
results: [
'anna', 'anat', 'anət', 'anna', 'tan', 'ənna'
]
})
});
});

View file

@ -1,4 +1,4 @@
import {stateReducer} from './reducer'; import {stateReducer} from './stateReducer';
it('default returns state unaltered', () => { it('default returns state unaltered', () => {
const state = {data: 'example'}; const state = {data: 'example'};

View file

@ -1,75 +0,0 @@
# LATL specification
## Feature Definition
## Rule Definition
ex.
```
(
`Unmotivated A to C`
A -> B / _
A -> C / _
``A becomes C in all environments with a intermediate state of B``
)
```
### Rule Body
#### Sound Definition
#### Change Definition
#### Environment Definition
##### Null Environment
Valid syntaxes:
```
A -> B ; no indicated environment
A -> B / _ ; environment indicated wth underscore
A -> B / . _ . ; environment indicated with underscore and placeholder dots
```
### Rule Metadata
#### Rule Title
#### Rule Description
## Language Primitives
## Data Structures
### Sets
Sets are collections of pointers to phones. The GLOBAL set contains all phones, making all other sets subsets of GLOBAL.
#### Global Set
[ GLOBAL ] is a shorthand for [ GLOBAL.SETS ]
#### Set Definition
Sets are defined with the set keyword followed by an equal sign and a set expression:
```
set SHORT_VOWELS = [ a, i, u ]
```
A single alias can be provided to the set during definition:
```
; the alias N can be used to refer to this set
set NASAL_PULMONIC_CONSONANTS, N = [ m, ɱ, n̼, n, ɳ, ɲ, ŋ, ɴ ]
```
Lists of sets can be defined using a comma followed by whitespace syntax
```
set PLOSIVES = [ p, t, k ],
FRICATIVES = [ f, s, x ],
LABIALIZED_PLOSIVES = { PLOSIVES yield [ X concat ʷ ] }
```
#### Set Usage
#### Set Operations
##### 'and' Operation
##### 'or' Operation
##### 'not' Operation
##### 'nor' Operation
##### 'in' Operation
##### 'yield' Operation
### Lexemes
#### Lexeme Operations
### Phone
For set of phones 'a', 'b', and 'ab':
```
GLOBAL ┬▻ <Key: a> ┬▻ <Key: b> ┬▻ { feature: <Boolean>, ... }
│ │ └▻ grapheme: <String: 'ab'>
│ └┬▻ { feature: <Boolean>, ... }
│ └▻ grapheme: <String: 'a'>
└┬▻ { feature: <Boolean>, ... }
└▻ grapheme: <String: 'b'>
```
#### Phone Operations
### Epochs

View file

@ -1,19 +0,0 @@
import { parser } from './parser';
export const codeGenerator = (latl) => {
const results = parser().feed(latl).results;
const nodeReader = (code, node) => {
if (node.length) {
return results.reduce(nodeReader, code)
}
if (!node) return code;
if (node.main) {
return nodeReader(code, node.main)
}
return code + node;
}
return nodeReader('', results)
}

View file

@ -1,120 +0,0 @@
// Generated automatically by nearley, version 2.19.1
// http://github.com/Hardmath123/nearley
(function () {
function id(x) { return x[0]; }
const { lexer } = require('./lexer.js');
const getTerminal = d => d ? d[0] : null;
const getAll = d => d.map((item, i) => ({ [i]: item }));
const flag = token => d => d.map(item => ({ [token]: item }))
const clearNull = d => d.filter(t => !!t && (t.length !== 1 || t[0])).map(t => t.length ? clearNull(t) : t);
const flagIndex = d => d.map((item, i) => ({[i]: item}))
const remove = _ => null;
const append = d => d.join('');
const constructSet = d => d.reduce((acc, t) => {
if (t && t.type === 'setIdentifier') acc.push({set: t});
if (t && t.length) acc[acc.length - 1].phones = t;
return acc;
}, []);
const pipe = (...funcs) => d => funcs.reduce((acc, func) => func(acc), d);
const objFromArr = d => d.reduce((obj, item) => ({ ...obj, ...item }), {});
var grammar = {
Lexer: lexer,
ParserRules: [
{"name": "main$ebnf$1", "symbols": []},
{"name": "main$ebnf$1$subexpression$1", "symbols": ["_", "statement"]},
{"name": "main$ebnf$1", "symbols": ["main$ebnf$1", "main$ebnf$1$subexpression$1"], "postprocess": function arrpush(d) {return d[0].concat([d[1]]);}},
{"name": "main", "symbols": ["main$ebnf$1", "_"], "postprocess": pipe(
clearNull,
// recursive call to fix repeat?
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
flag('main'),
getTerminal,
) },
{"name": "_$ebnf$1$subexpression$1", "symbols": [(lexer.has("whiteSpace") ? {type: "whiteSpace"} : whiteSpace)]},
{"name": "_$ebnf$1", "symbols": ["_$ebnf$1$subexpression$1"], "postprocess": id},
{"name": "_$ebnf$1", "symbols": [], "postprocess": function(d) {return null;}},
{"name": "_", "symbols": ["_$ebnf$1"], "postprocess": remove},
{"name": "__", "symbols": [(lexer.has("whiteSpace") ? {type: "whiteSpace"} : whiteSpace)], "postprocess": remove},
{"name": "equal", "symbols": [(lexer.has("equal") ? {type: "equal"} : equal)], "postprocess": remove},
{"name": "statement", "symbols": ["comment"]},
{"name": "statement", "symbols": ["definition"], "postprocess": pipe(
d => d.flatMap(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
// recursive call to fit repeat?
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
// may split from other definition statements
d => d.map(t => t && t.length > 1 ? ({ type: 'set', ...objFromArr(t) }) : null)
) },
{"name": "comment", "symbols": [(lexer.has("comment") ? {type: "comment"} : comment)], "postprocess": pipe(getTerminal, remove)},
{"name": "definition$ebnf$1", "symbols": []},
{"name": "definition$ebnf$1$subexpression$1", "symbols": ["setDefinition", (lexer.has("comma") ? {type: "comma"} : comma), "__"]},
{"name": "definition$ebnf$1", "symbols": ["definition$ebnf$1", "definition$ebnf$1$subexpression$1"], "postprocess": function arrpush(d) {return d[0].concat([d[1]]);}},
{"name": "definition", "symbols": [(lexer.has("kwSet") ? {type: "kwSet"} : kwSet), "__", "definition$ebnf$1", "setDefinition"], "postprocess": pipe(
// not yet sure why this call is required twice
d => d.map(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
d => d.map(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
d => d.map(u => u && u.length ? u.map(v => v.length ? v.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet')[0] : v) : u),
clearNull,
) },
{"name": "setDefinition$ebnf$1$subexpression$1", "symbols": ["setAlias"]},
{"name": "setDefinition$ebnf$1", "symbols": ["setDefinition$ebnf$1$subexpression$1"], "postprocess": id},
{"name": "setDefinition$ebnf$1", "symbols": [], "postprocess": function(d) {return null;}},
{"name": "setDefinition", "symbols": [(lexer.has("setIdentifier") ? {type: "setIdentifier"} : setIdentifier), "setDefinition$ebnf$1", "__", "equal", "__", "setExpression"], "postprocess":
pipe(
d => d.filter(t => !!t && t.length !== 0),
d => d.map(u => u && u.length ? u.map(t => t && t.length ? t.filter(v => v && v.type !== 'comma') : t) : u),
d => d.map(t => t.type === 'setIdentifier' ? { setIdentifier: t.toString() } : t),
d => d.map(t => t && t.length && t[0].hasOwnProperty('setExpression') ? t[0] : t),
d => d.map(t => t.length ?
// pretty ugly ([ { type: 'aias', alias: [ string ] }] ) => { setAlias: str }
{ setAlias: t.reduce((aliases, token) => token && token.type === 'alias' ? [...aliases, ...token.alias] : aliases, [])[0] }
: t),
)
},
{"name": "setExpression", "symbols": [(lexer.has("openSquareBracket") ? {type: "openSquareBracket"} : openSquareBracket), "_", "phoneList", "_", (lexer.has("closeSquareBracket") ? {type: "closeSquareBracket"} : closeSquareBracket)]},
{"name": "setExpression$ebnf$1$subexpression$1", "symbols": ["setOperation"]},
{"name": "setExpression$ebnf$1", "symbols": ["setExpression$ebnf$1$subexpression$1"], "postprocess": id},
{"name": "setExpression$ebnf$1", "symbols": [], "postprocess": function(d) {return null;}},
{"name": "setExpression", "symbols": [(lexer.has("openCurlyBracket") ? {type: "openCurlyBracket"} : openCurlyBracket), "_", "setExpression$ebnf$1", "_", (lexer.has("closeCurlyBracket") ? {type: "closeCurlyBracket"} : closeCurlyBracket)], "postprocess":
pipe(
// filters commas and whitespace
d => d.filter(t => t && t.length),
d => d.map(t => t.map(u => u[0])),
flag('setExpression')
) },
{"name": "setAlias", "symbols": [(lexer.has("comma") ? {type: "comma"} : comma), "_", (lexer.has("setIdentifier") ? {type: "setIdentifier"} : setIdentifier)], "postprocess": pipe(
d => d && d.length ? d.filter(t => !!t) : d,
d => d.map(t => t.type === 'setIdentifier' ? t.toString() : null),
d => d.filter(t => !!t),
d => ({type: 'alias', alias: d }),
) },
{"name": "phoneList$ebnf$1", "symbols": []},
{"name": "phoneList$ebnf$1$subexpression$1$ebnf$1", "symbols": []},
{"name": "phoneList$ebnf$1$subexpression$1$ebnf$1$subexpression$1", "symbols": [(lexer.has("comma") ? {type: "comma"} : comma), "_"]},
{"name": "phoneList$ebnf$1$subexpression$1$ebnf$1", "symbols": ["phoneList$ebnf$1$subexpression$1$ebnf$1", "phoneList$ebnf$1$subexpression$1$ebnf$1$subexpression$1"], "postprocess": function arrpush(d) {return d[0].concat([d[1]]);}},
{"name": "phoneList$ebnf$1$subexpression$1", "symbols": [(lexer.has("phone") ? {type: "phone"} : phone), "phoneList$ebnf$1$subexpression$1$ebnf$1"]},
{"name": "phoneList$ebnf$1", "symbols": ["phoneList$ebnf$1", "phoneList$ebnf$1$subexpression$1"], "postprocess": function arrpush(d) {return d[0].concat([d[1]]);}},
{"name": "phoneList", "symbols": ["phoneList$ebnf$1"], "postprocess":
pipe(
d => d ? d[0].map(t => t.filter(u => u.type === 'phone').map(u => u.toString())) : d
)
},
{"name": "setOperation", "symbols": ["orOperation"]},
{"name": "setOperation", "symbols": [(lexer.has("identifier") ? {type: "identifier"} : identifier)], "postprocess": pipe(
d => d.type ? d : ({ identifier: d.toString(), type: 'identifier' })
)},
{"name": "orOperation", "symbols": ["_", "setOperation", "__", (lexer.has("kwSetOr") ? {type: "kwSetOr"} : kwSetOr), "__", "setOperation", "_"], "postprocess": pipe(
d => d.filter(d => !!d),
d => ({ type: 'operator', operator: 'or', operands: [ d[0], d[2] ] }),
) }
]
, ParserStart: "main"
}
if (typeof module !== 'undefined'&& typeof module.exports !== 'undefined') {
module.exports = grammar;
} else {
window.grammar = grammar;
}
})();

View file

@ -1,109 +0,0 @@
@{%
const { lexer } = require('./lexer.js');
const getTerminal = d => d ? d[0] : null;
const getAll = d => d.map((item, i) => ({ [i]: item }));
const flag = token => d => d.map(item => ({ [token]: item }))
const clearNull = d => d.filter(t => !!t && (t.length !== 1 || t[0])).map(t => t.length ? clearNull(t) : t);
const flagIndex = d => d.map((item, i) => ({[i]: item}))
const remove = _ => null;
const append = d => d.join('');
const constructSet = d => d.reduce((acc, t) => {
if (t && t.type === 'setIdentifier') acc.push({set: t});
if (t && t.length) acc[acc.length - 1].phones = t;
return acc;
}, []);
const pipe = (...funcs) => d => funcs.reduce((acc, func) => func(acc), d);
const objFromArr = d => d.reduce((obj, item) => ({ ...obj, ...item }), {});
%}
@lexer lexer
main -> (_ statement):* _
{% pipe(
clearNull,
// recursive call to fix repeat?
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
flag('main'),
getTerminal,
) %}
_ -> (%whiteSpace):?
{% remove %}
__ -> %whiteSpace
{% remove %}
equal -> %equal
{% remove %}
statement -> comment | definition
{% pipe(
d => d.flatMap(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
// recursive call to fit repeat?
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
d => d.map(t => t && t.length === 1 && t[0] ? t[0] : t),
// may split from other definition statements
d => d.map(t => t && t.length > 1 ? ({ type: 'set', ...objFromArr(t) }) : null)
) %}
comment -> %comment
{% pipe(getTerminal, remove) %}
# SETS
definition -> %kwSet __ (setDefinition %comma __):* setDefinition
{% pipe(
// not yet sure why this call is required twice
d => d.map(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
d => d.map(u => u && u.length ? u.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet') : u),
d => d.map(u => u && u.length ? u.map(v => v.length ? v.filter(t => t && t.type !== 'comma' && t.type !== 'kwSet')[0] : v) : u),
clearNull,
) %}
setDefinition -> %setIdentifier (setAlias):? __ equal __ setExpression
{%
pipe(
d => d.filter(t => !!t && t.length !== 0),
d => d.map(u => u && u.length ? u.map(t => t && t.length ? t.filter(v => v && v.type !== 'comma') : t) : u),
d => d.map(t => t.type === 'setIdentifier' ? { setIdentifier: t.toString() } : t),
d => d.map(t => t && t.length && t[0].hasOwnProperty('setExpression') ? t[0] : t),
d => d.map(t => t.length ?
// pretty ugly ([ { type: 'aias', alias: [ string ] }] ) => { setAlias: str }
{ setAlias: t.reduce((aliases, token) => token && token.type === 'alias' ? [...aliases, ...token.alias] : aliases, [])[0] }
: t),
)
%}
setExpression -> %openSquareBracket _ phoneList _ %closeSquareBracket
| %openCurlyBracket _ (setOperation):? _ %closeCurlyBracket
{%
pipe(
// filters commas and whitespace
d => d.filter(t => t && t.length),
d => d.map(t => t.map(u => u[0])),
flag('setExpression')
) %}
setAlias -> %comma _ %setIdentifier
{% pipe(
d => d && d.length ? d.filter(t => !!t) : d,
d => d.map(t => t.type === 'setIdentifier' ? t.toString() : null),
d => d.filter(t => !!t),
d => ({type: 'alias', alias: d }),
) %}
phoneList -> (%phone (%comma _):* ):*
{%
pipe(
d => d ? d[0].map(t => t.filter(u => u.type === 'phone').map(u => u.toString())) : d
)
%}
setOperation -> orOperation
| %identifier
{% pipe(
d => d.type ? d : ({ identifier: d.toString(), type: 'identifier' })
)%}
orOperation -> _ setOperation __ %kwSetOr __ setOperation _
{% pipe(
d => d.filter(d => !!d),
d => ({ type: 'operator', operator: 'or', operands: [ d[0], d[2] ] }),
) %}

View file

@ -1,124 +0,0 @@
const moo = require("moo");
const lexer = moo.states({
main: {
comment: /;.*$/,
star: { match: /\*/, push: "epoch" },
slash: { match: /\//, push: "lexicon" },
// change so that identifiers are always upper, keywords are always lower, phones are always lower
kwSet: {
match: "set",
type: moo.keywords({ kwSet: "set " }),
push: "setDefinition",
},
identifier: { match: /[A-Za-z]+[\u00c0-\u03FFA-Za-z0-9\\-\\_]*/ },
openBracket: { match: /\[/, push: "feature" },
whiteSpace: { match: /\s+/, lineBreaks: true },
newLine: { match: /\n+/, lineBreaks: true },
},
epoch: {
identifier: {
match: /[A-Za-z]+[\u00c0-\u03FFA-Za-z0-9\\-\\_]*/,
push: "rule",
},
openParen: { match: /\(/, push: "ruleDefinition" },
pipe: { match: /\|/, pop: true },
greaterThan: /\>/,
arrow: /\-\>/,
hash: /#/,
slash: /\//,
dot: /\./,
underscore: /\_/,
newLine: { match: /\n/, lineBreaks: true },
},
ruleDefinition: {
doubleTick: { match: /``/, push: "ruleName" },
singleTick: { match: /`/, push: "ruleDescription" },
// push rule
closeParen: { match: /\)/, pop: true },
newLine: { match: /\n/, lineBreaks: true },
},
ruleName: {
ruleName: { match: /.+(?=``)/ },
doubleTick: { match: /``/, pop: true },
},
ruleDescription: {
ruleDescription: { match: /.+(?=`)/ },
singleTick: { match: /`/, pop: true },
},
rule: {
openSquareBracket: { match: /\[/, push: "ruleFeature" },
// whiteSpace: { match: /\s/ },
newLine: { match: /\n/, pop: true, lineBreaks: true },
},
ruleFeature: {
ruleFeature: { match: /[A-Za-z]+[\u00c0-\u03FFA-Za-z0-9\\-\\_]*/ },
closeBracket: { match: /\]/, pop: true },
newLine: { match: /\n/, lineBreaks: true },
},
lexicon: {
slash: { match: /\//, pop: true },
newLine: { match: /\n/, lineBreaks: true },
},
feature: {
closeBracket: { match: /\]/, pop: true },
positiveAssignment: /\+=/,
negativeAssignment: /\-=/,
newLine: { match: /\n/, lineBreaks: true },
},
setDefinition: {
comment: /;.*$/,
setIdentifier: { match: /[A-Z]+[A-Z_]*/ },
openCurlyBracket: { match: /\{/, push: "setOperation" },
equal: /=/,
openSquareBracket: /\[/,
phone: /[\u00c0-\u03FFa-z]+/,
closeSquareBracket: { match: /\]/ },
comma: { match: /,/, push: "commaOperation" },
whiteSpace: { match: /[\t ]+/ },
newLine: { match: /\n/, pop: true, lineBreaks: true },
},
setOperation: {
closeCurlyBracket: { match: /\}/, pop: true },
// ! restrict identifiers
keyword: {
match: ["not", "and", "or", "nor", "in", "yield", "concat", "dissoc"],
type: moo.keywords({
kwSetNot: "not",
kwSetAnd: "and",
kwSetOr: "or",
kwSetNor: "nor",
kwSetIn: "in",
kwSetYield: "yield",
kwSetConcat: "concat",
kwSetDissoc: "dissoc",
}),
},
identifier: /[A-Z]+[A-Z_]+/,
whiteSpace: { match: /\s+/, lineBreaks: true },
openSquareBracket: /\[/,
closeSquareBracket: /\]/,
identifier: /[A-Z]+[A-Z_]*/,
phone: /[\u00c0-\u03FFa-z]+/,
},
commaOperation: {
// if comma is detected during a definition, the commaOperation consumes all white space and pops back to definition
// this prevents popping back to main
comment: /\s*;.*$/,
whiteSpace: { match: /\s+/, lineBreaks: true, pop: true },
newLine: { match: /\n/, lineBreaks: true, pop: true },
},
});
module.exports = { lexer };

View file

@ -1,4 +0,0 @@
const nearley = require("nearley");
const grammar = require("./grammar.js");
export const parser = () => new nearley.Parser(nearley.Grammar.fromCompiled(grammar));

View file

@ -1,810 +0,0 @@
export const assertionData = {
simpleComment: {
latl: `; comment`,
tokens: [{ type: "comment", value: "; comment" }],
AST: {
main: [],
},
code: "",
},
simpleSetDefinition: {
latl: `set NASAL_PULMONIC_CONSONANTS = [ m̥, m, ɱ ]`,
tokens: [
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "NASAL_PULMONIC_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m̥" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɱ" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
],
AST: {
main: [
{
type: "set",
setIdentifier: "NASAL_PULMONIC_CONSONANTS",
setExpression: ["m̥", "m", "ɱ"],
},
],
},
code: "",
},
commaSetDefinition: {
latl: `
set NASAL_PULMONIC_CONSONANTS = [ m̥, m, ɱ, n̼, n̥, n, ɳ̊, ɳ, ɲ̊, ɲ, ŋ, ̊ŋ, ɴ ],
STOP_PULMONIC_CONSONANTS = [ p, b, p̪, b̪, t̼, d̼, t, d, ʈ, ɖ, c, ɟ, k, ɡ, q, ɢ, ʡ, ʔ ]`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "NASAL_PULMONIC_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m̥" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɱ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n̼" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n̥" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɳ̊" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɳ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɲ̊" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɲ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ŋ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "̊ŋ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɴ" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "STOP_PULMONIC_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "p" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "b" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "p̪" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "b̪" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "t̼" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "d̼" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "t" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "d" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ʈ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɖ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "c" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɟ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "k" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɡ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "q" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɢ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ʡ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ʔ" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
],
AST: {
main: [
{
type: "set",
setIdentifier: "NASAL_PULMONIC_CONSONANTS",
setExpression: [
"m̥",
"m",
"ɱ",
"n̼",
"n̥",
"n",
"ɳ̊",
"ɳ",
"ɲ̊",
"ɲ",
"ŋ",
"̊ŋ",
"ɴ",
],
},
{
type: "set",
setIdentifier: "STOP_PULMONIC_CONSONANTS",
setExpression: [
"p",
"b",
"p̪",
"b̪",
"t̼",
"d̼",
"t",
"d",
"ʈ",
"ɖ",
"c",
"ɟ",
"k",
"ɡ",
"q",
"ɢ",
"ʡ",
"ʔ",
],
},
],
},
},
setAliasDefinition: {
latl: `
set NASAL_PULMONIC_CONSONANTS, N = [ m̥, m, ɱ, n̼, n̥, n, ɳ̊, ɳ, ɲ̊, ɲ, ŋ, ̊ŋ, ɴ ]`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "NASAL_PULMONIC_CONSONANTS" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "N" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m̥" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "m" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɱ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n̼" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n̥" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "n" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɳ̊" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɳ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɲ̊" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɲ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ŋ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "̊ŋ" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "ɴ" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
],
AST: {
main: [
{
type: "set",
setIdentifier: "NASAL_PULMONIC_CONSONANTS",
setAlias: "N",
setExpression: [
"m̥",
"m",
"ɱ",
"n̼",
"n̥",
"n",
"ɳ̊",
"ɳ",
"ɲ̊",
"ɲ",
"ŋ",
"̊ŋ",
"ɴ",
],
},
],
},
},
setDefinitionJoin: {
latl: `
set CLICK_CONSONANTS = { TENUIS_CLICK_CONSONANTS or VOICED_CLICK_CONSONANTS }`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "TENUIS_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetOr", value: "or" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "VOICED_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
],
AST: {
main: [
{
type: "set",
setIdentifier: "CLICK_CONSONANTS",
setExpression: [
{
type: "operator",
operator: "or",
operands: [
{
type: "identifier",
identifier: "TENUIS_CLICK_CONSONANTS",
},
{
type: "identifier",
identifier: "VOICED_CLICK_CONSONANTS",
},
],
},
],
},
],
},
},
setDefinitionMultiJoin: {
latl: `
set CLICK_CONSONANTS = { TENUIS_CLICK_CONSONANTS or VOICED_CLICK_CONSONANTS
or NASAL_CLICK_CONSONANTS or L_CLICK_CONSONANTS
}`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "TENUIS_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetOr", value: "or" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "VOICED_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: "\n " },
{ type: "kwSetOr", value: "or" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "NASAL_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetOr", value: "or" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "L_CLICK_CONSONANTS" },
{ type: "whiteSpace", value: " \n " },
{ type: "closeCurlyBracket", value: "}" },
],
AST: {
main: [
{
type: "set",
setIdentifier: "CLICK_CONSONANTS",
setExpression: [
{
type: "operator",
operator: "or ",
operands: [
{
type: "identifier",
identifier: "TENUIS_CLICK_CONSONANTS",
},
{
type: "operator",
operator: "or",
operands: [
{
type: "identifier",
identifier: "VOICED_CLICK_CONSONANTS",
},
{
type: "operator",
operator: "or",
operands: [
{
type: "identifier",
identifier: "NASAL_CLICK_CONSONANTS",
},
{
type: "identifier",
operands: "L_CLICK_CONSONANTS",
},
],
},
],
},
],
},
],
},
],
},
},
setDefinitionYield: {
latl: `
set NASAL_VOWELS = { [ V ] in ORAL_VOWELS yield [ Ṽ ] },
SHORT_NASAL_VOWELS = { [ Vː ] in NASAL_VOWELS yield [ V ]ː },
LONG_NASAL_VOWELS = { [ Vː ] in NASAL_VOWELS }`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "NASAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "V" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetIn", value: "in" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "ORAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "V" },
{ type: "phone", value: "̃" }, // test display for COMBINING TILDE OVERLAY is deceiving
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SHORT_NASAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "V" },
{ type: "phone", value: "ː" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetIn", value: "in" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "NASAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "V" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "phone", value: "ː" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "LONG_NASAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "V" },
{ type: "phone", value: "ː" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetIn", value: "in" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "NASAL_VOWELS" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
],
},
setOperationsJoin: {
latl: `
; ---- set join operations non-mutable!
set SET_C = { SET_A not SET_B }, ; left anti join
SET_D = { SET_A and SET_B }, ; inner join
SET_E = { SET_A or SET_B }, ; full outer join
SET_F = { not SET_A }, ; = { GLOBAL not SET_A }
SET_G = { not SET_A nor SET_B } ; = { GLOBAL not { SET_A or SET_B } }`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{ type: "comment", value: "; ---- set join operations non-mutable! " },
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "SET_C" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetNot", value: "not" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_B" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "comment", value: " ; left anti join" },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_D" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetAnd", value: "and" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_B" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "comment", value: " ; inner join" },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_E" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetOr", value: "or" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_B" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "comment", value: " ; full outer join" },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_F" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetNot", value: "not" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "comment", value: " ; = { GLOBAL not SET_A }" },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_G" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetNot", value: "not" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetNor", value: "nor" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_B" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "whiteSpace", value: " " },
{ type: "comment", value: "; = { GLOBAL not { SET_A or SET_B } }" },
],
},
setOperations: {
latl: `
; ---- set character operations - non-mutable!
set SET_B = { [ Xy ] in SET_A }, ; FILTER: where X is any character and y is a filtering character
SET_C = { SET_A yield [ Xy ] }, ; CONCATENATE: performs transformation with (prepended or) appended character
SET_D = { SET_A yield [ X concat y ] },
SET_E = { SET_A yield [ y concat X ] },
SET_F = { SET_A yield y[ X ] }, ; DISSOCIATE: performs transformation removing prepended (or appended) character
SET_G = { SET_A yield y dissoc [ X ] },
SET_H = { SET_A yield [ X ] dissoc y },
SET_I = { [ Xy ] in SET_A yield [ X ]y } ; combined FILTER and DISSOCIATE`,
tokens: [
{ type: "whiteSpace", value: "\n" },
{
type: "comment",
value: "; ---- set character operations - non-mutable!",
},
{ type: "whiteSpace", value: "\n" },
{ type: "kwSet", value: "set" },
{ type: "whiteSpace", value: " " },
{ type: "setIdentifier", value: "SET_B" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetIn", value: "in" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{
type: "comment",
value:
" ; FILTER: where X is any character and y is a filtering character",
},
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_C" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{
type: "comment",
value:
" ; CONCATENATE: performs transformation with (prepended or) appended character",
},
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_D" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetConcat", value: "concat" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_E" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetConcat", value: "concat" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_F" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "y" },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{
type: "comment",
value:
" ; DISSOCIATE: performs transformation removing prepended (or appended) character",
},
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_G" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetDissoc", value: "dissoc" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_H" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetDissoc", value: "dissoc" },
{ type: "whiteSpace", value: " " },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "comma", value: "," },
{ type: "whiteSpace", value: "\n " },
{ type: "setIdentifier", value: "SET_I" },
{ type: "whiteSpace", value: " " },
{ type: "equal", value: "=" },
{ type: "whiteSpace", value: " " },
{ type: "openCurlyBracket", value: "{" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetIn", value: "in" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "SET_A" },
{ type: "whiteSpace", value: " " },
{ type: "kwSetYield", value: "yield" },
{ type: "whiteSpace", value: " " },
{ type: "openSquareBracket", value: "[" },
{ type: "whiteSpace", value: " " },
{ type: "identifier", value: "X" },
{ type: "whiteSpace", value: " " },
{ type: "closeSquareBracket", value: "]" },
{ type: "phone", value: "y" },
{ type: "whiteSpace", value: " " },
{ type: "closeCurlyBracket", value: "}" },
{ type: "whiteSpace", value: " " },
{ type: "comment", value: "; combined FILTER and DISSOCIATE" },
],
},
};

View file

@ -1,10 +0,0 @@
import { assertionData } from './assertionData';
import { codeGenerator } from '../codeGenerator';
describe('codeGenerator', () => {
it('parses simple comment', () => {
const { latl, code } = assertionData.simpleComment;
const generatedCode = codeGenerator(latl);
expect(generatedCode).toEqual(code);
});
})

View file

@ -1,71 +0,0 @@
import { lexer } from '../lexer';
import { assertionData } from './assertionData';
describe('lexer', () => {
const getToken = obj => obj ? formatToken(obj) : null;
const formatToken = obj => ({ type: obj.type, value: obj.value });
const getStream = latl => {
lexer.reset(latl);
let token = getToken(lexer.next());
let stream = [];
do {
stream = [...stream, token]
token = getToken(lexer.next());
} while (token);
return stream;
}
it('lexes simple comment', () => {
const { latl, tokens } = assertionData.simpleComment;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
// it('lexes simple * and identifier', () => {
// lexer.reset('*proto');
// const stream = [ getToken(lexer.next()), getToken(lexer.next()) ];
// expect(stream).toStrictEqual([ { type: 'star', value: '*' }, { type: 'identifier', value: 'proto' } ]);
// })
it('lexes set and identifier', () => {
const { latl, tokens } = assertionData.simpleSetDefinition;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
})
it('lexes multiple set definitions with comma operator', () => {
const { latl, tokens } = assertionData.commaSetDefinition;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
it('lexes set definition with alias', () => {
const { latl, tokens } = assertionData.setAliasDefinition;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
it('lexes set definition with set join', () => {
const { latl, tokens } = assertionData.setDefinitionJoin;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
it('lexes set definition with yield operation', () => {
const { latl, tokens } = assertionData.setDefinitionYield;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
it('lexes all set join operations', () => {
const { latl, tokens } = assertionData.setOperationsJoin;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
});
it('lexes set filter, concat, and dissoc operations', () => {
const { latl, tokens } = assertionData.setOperations;
const stream = getStream(latl);
expect(stream).toStrictEqual(tokens);
})
})

View file

@ -1,180 +0,0 @@
import { lexer } from "../lexer";
import { parser } from "../parser";
import { assertionData } from "./assertionData";
describe("parser", () => {
it("parses simple comment", () => {
const { latl, AST } = assertionData.simpleComment;
const feedResults = parser().feed(latl).results;
expect(feedResults.length).toBe(1);
expect(feedResults[0]).toStrictEqual(AST);
});
it("parses simple set definition", () => {
const { latl, AST } = assertionData.simpleSetDefinition;
const feedResults = parser().feed(latl).results;
expect(feedResults.length).toBe(1);
expect(feedResults[0]).toStrictEqual(AST);
});
it("parses multiple set definitions with comma operator", () => {
const { latl, AST } = assertionData.commaSetDefinition;
const feedResults = parser().feed(latl).results;
expect(feedResults.length).toBe(1);
expect(feedResults[0]).toStrictEqual(AST);
});
it("lexes set definition with alias", () => {
const { latl, AST } = assertionData.setAliasDefinition;
const feedResults = parser().feed(latl).results;
expect(feedResults[0]).toStrictEqual(AST);
});
it.skip("lexes set definition with set join", () => {
const { latl, AST } = assertionData.setDefinitionJoin;
const feedResults = parser().feed(latl).results;
expect(feedResults[0]).toStrictEqual(AST);
});
it.todo(
"lexes set definition with yield operation"
// , () => {
// const { latl, tokens } = assertionData.setDefinitionYield;
// const stream = getStream(latl);
// expect(stream).toStrictEqual(tokens);
// }
);
it.todo(
"lexes all set join operations"
// , () => {
// const { latl, tokens } = assertionData.setOperationsJoin;
// const stream = getStream(latl);
// expect(stream).toStrictEqual(tokens);
// }
);
it.todo(
"lexes set filter, concat, and dissoc operations"
// , () => {
// const { latl, tokens } = assertionData.setOperations;
// const stream = getStream(latl);
// expect(stream).toStrictEqual(tokens);
// }
);
});
// {
// "set":
// [
// [
// [
// {
// "col": 5,
// "line": 2,
// "lineBreaks": 0,
// "offset": 5,
// "text": "NASAL_PULMONIC_CONSONANTS",
// "toString": [tokenToString],
// "type": "setIdentifier",
// "value": "NASAL_PULMONIC_CONSONANTS",
// },
// null,
// {
// "col": 45,
// "line": 2,
// "lineBreaks": 0,
// "offset": 45,
// "text": "=",
// "toString": [tokenToString],
// "type": "equal",
// "value": "=",
// },
// null,
// [
// [
// {
// "col": 49,
// "line": 2,
// "lineBreaks": 0,
// "offset": 49,
// "text": "m̥",
// "toString": [tokenToString],
// "type": "phone",
// "value": "m̥",
// },
// {
// "col": 91,
// "line": 2,
// "lineBreaks": 0,
// "offset": 91,
// "text": "ɴ",
// "toString": [tokenToString],
// "type": "phone",
// "value": "ɴ",
// },
// ],
// ],
// {
// "col": 94,
// "line": 2,
// "lineBreaks": 0,
// "offset": 94,
// "text": ",",
// "toString": [tokenToString],
// "type": "comma",
// "value": ",",
// },
// null,
// ],
// ],
// - "setIdentifier": "STOP_PULMONIC_CONSONANTS",
// {
// "col": 5,
// "line": 3,
// "lineBreaks": 0,
// "offset": 100,
// "text": "STOP_PULMONIC_CONSONANTS",
// "toString": [tokenToString],
// "type": "setIdentifier",
// "value": "STOP_PULMONIC_CONSONANTS",
// },
// null,
// {
// "col": 45,
// "line": 3,
// "lineBreaks": 0,
// "offset": 140,
// "text": "=",
// "toString": [tokenToString],
// "type": "equal",
// "value": "=",
// },
// null,
// [
// [
// {
// "col": 49,
// "line": 3,
// "lineBreaks": 0,
// "offset": 144,
// "text": "p",
// "toString": [tokenToString],
// "type": "phone",
// "value": "p",
// },
// {
// "col": 104,
// "line": 3,
// "lineBreaks": 0,
// "offset": 199,
// "text": "ʔ",
// "toString": [tokenToString],
// "type": "phone",
// "value": "ʔ",
// },
// ],
// ],
// ],
// "token": "kwSet",
// }