Improving Rikaitan's features for the language(s) you are interested in is pretty simple, and a great way to help yourself and others. This guide will help you get started.
If your language is not already available in the Language dropdown, here is how you can add it with just a few lines. As an example, we'll use PR #913, where a first-time contributor added Dutch.
// language-descriptors.js
{
iso: 'nl',
iso639_3: 'nld',
name: 'Dutch',
exampleText: 'lezen',
textPreprocessors: capitalizationPreprocessors
}
- Look up the ISO 639-1 and ISO 639-3 codes for your language. If it is a rarer language it might not have a ISO-639-1 code - if that's the case, use ISO 639-3 for both
iso
andiso639_3
. - Place your language in the
languageDescriptors
array inlanguage-descriptors.js
. The languages are sorted alphabetically by ISO code. - The example text is usually some form of the verb "to read" in your language, but it can be any example you feel is good. This will be shown in the preview popup on the settings page.
- If your language uses the Latin or Cyrillic script, or another script with capitalization, you will want to use the
capitalizationPreprocessors
. We'll cover this part in more detail a bit later. ThetextPreprocessors
field can also be left out.
When in doubt, look at the other languages in the file for ideas. The same applies to language-descriptors.d.ts
:
// language-descriptors.d.ts
nl: {
pre: CapitalizationPreprocessors;
}
This is just for some type safety. The first key is the ISO code. Most languages will then only have a pre
key (the other one is post
), and its value is the type of text preprocessors you used in language-descriptors.js
. Use the TypeScript operator &
as needed. If you didn't use any text preprocessors, you can set the value to Record<string, never>
.
That's it! Your language should now be selectable from the dropdown, and may work perfectly fine already. If you don't already have a dictionary to test with, check out Dictionaries. For more advanced features, read on.
You should first have the repo set up locally according to the instructions in the contributing guidelines.
A language descriptor in language-descriptors.js
has several optional fields for more advanced features. We've already mentioned textPreprocessors
, but there are also languageTransforms
, textPostprocessors
, isTextLookupWorthy
, and readingNormalizer
. Let's go through them (see also the full type definition in language-descriptors.d.ts
).
The scanned text may not exactly match the word in the dictionary. For example, an English dictionary will likely contain the word "read", but the text may contain "Read" or "READ". To handle cases like this, we use text preprocessors.
// from language.d.ts
export type TextProcessor<T = unknown> = {
name: string;
description: string;
options: TextProcessorOptions<T>;
process: TextProcessorFunction<T>;
};
Every text preprocessor has:
- A
name
anddescription
- An array of
options
, most commonly just[false, true]
, that control the behavior of theprocess
function. - A
process
function that takes a string and a setting and returns a string
Here are the CapitalizationPreprocessors
used in the Dutch example:
/** @type {import('language').TextProcessor<boolean>} */
export const decapitalize = {
name: "Decapitalize text",
description: "CAPITALIZED TEXT → capitalized text",
options: basicTextProcessorOptions, // [false, true]
process: (str, setting) => (setting ? str.toLowerCase() : str),
};
/** @type {import('language').TextProcessor<boolean>} */
export const capitalizeFirstLetter = {
name: "Capitalize first letter",
description: "lowercase text → Lowercase text",
options: basicTextProcessorOptions, // [false, true]
process: (str, setting) =>
setting ? str.charAt(0).toUpperCase() + str.slice(1) : str,
};
When applying preprocessors, each combination will be separately applied and looked up. Since each of these two preprocessors has two options, there are 2*2=4 possible combinations. For the input string reaD
, the following strings will be looked up:
reaD
(both preprocessors off)ReaD
(onlycapitalizeFirstLetter
on)read
(onlydecapitalize
on)Read
(both preprocessors on)
Note that the order of text processors can matter. Had we put capitalizeFirstLetter before decapitalize, the 4th string would be read
instead of Read
.
A letter or a group of letters may have multiple variants in a language. For example, in German, "ß" can be written as "ss" and vice versa, or in Japanese every kana has a hiragana and a katakana variant. To handle this, we use a bidirectional conversion preprocessor.
// from german-text-preprocessors.js
/** @type {import('language').BidirectionalConversionPreprocessor} */
export const eszettPreprocessor = {
name: 'Convert "ß" to "ss"',
description: "ß → ss, ẞ → SS and vice versa",
options: ["off", "direct", "inverse"],
process: (str, setting) => {
switch (setting) {
case "off":
return str;
case "direct":
return str.replace(/ẞ/g, "SS").replace(/ß/g, "ss");
case "inverse":
return str.replace(/SS/g, "ẞ").replace(/ss/g, "ß");
}
},
};
These have three options: off, direct, and inverse, and the process
function must handle each of them.
In some cases (e.g. German umlauts), diacritics are near-ubiquitous. However some languages (such as Latin, Arabic etc) do not commonly use diacritics, but only in specific kinds of texts (e.g dictionaries, texts for children or learners). In these cases, the dictionaries that Rikaitan uses will likely not contain diacritics, but the text may contain them. To handle this, we use a diacritics removal preprocessor.
This kind of text processing is to a degree interdependent with the dictionaries available for the language.
Deinflection is the process of converting a word to its base or dictionary form. For example, "running" would be deinflected to "run". This is useful for finding the word in the dictionary, as well as helping the user understand the grammar (morphology) of the language.
These grammatical rules are located in files such as english-transforms.js
.
Not all the grammatical rules of a language can or need to be implemented in the transforms file. Even a little bit goes a long way, and you can always add more rules later. For every couple rules you add, write some tests in the respective file (see the writing tests section below). This will help you verify that your rules are correct, and make sure nothing is accidentally broken along the way.
Transforms files should export a LanguageTransformDescriptor
, which is then imported in language-descriptors.js
.
// from language-transformer.d.ts
export type LanguageTransformDescriptor<TCondition extends string = string> = {
language: string;
conditions: ConditionMapObject<TCondition>;
transforms: TransformMapObject<TCondition>;
};
export type ConditionMapObject<TCondition extends string> = {
[type in TCondition]: Condition;
};
export type TransformMapObject<TCondition> = {
[name: string]: Transform<TCondition>;
};
language
is the ISO code of the languageconditions
are an object containing parts of speech and grammatical forms that are used to check which deinflections make sense. They are referenced by the deinflection rules.transforms
are the actual deinflection rulesTCondition
is an optional generic parameter that can be passed toLanguageTransformDescriptor
. You can learn more about it at the end of this section.
Let's try and write a bit of deinflection for English, from scratch.
// english-transforms.js
import { suffixInflection } from "../language-transforms.js";
export const englishTransforms = {
language: "en",
conditions: {},
transforms: {
plural: {
name: "plural",
description: "Plural form of a noun",
rules: [suffixInflection("s", "", [], [])],
},
},
};
This is a simple example for English, where the only deinflection rule is to remove the "s" from the end of a noun to get the singular form. The suffixInflection
function is a helper that creates a deinflection rule for a suffix. It takes the suffix to remove, what to replace it with, and two more parameters for conditions, which we will look at next. The suffixInflection
is the most common type of deinflection rule across languages. The inner plural
is the displayed description while looking up, and the outer plural
is a name only to be referenced internally within the file.
For the input string "cats", the following strings will be looked up:
cats
(no deinflection)cat
(deinflected by theplural
rule)
If the dictionary contains an entry for cat
, it will successfully match the 2nd looked up string, (as shown in the image). Note the 🧩 symbol and the plural
rule.
However, this rule will also match the word "reads", and show the verb "read" from the dictionary, marked as being plural
. This makes no sense, and we can use conditions to prevent it. Let's add a condition and use it in the rule.
conditions: {
n: {
name: 'Noun',
isDictionaryForm: true,
},
},
transforms: {
"plural": {
name: "plural",
description: "Plural form of a noun",
rules: [
suffixInflection("s", "", [], ["n"])
],
},
},
Now, only dictionary entries marked with the same "n" condition will be eligible for matching the plural
rule. The verb "read" should be marked as "v" in the dictionary, and will no longer be matched by the plural
rule. The entries in the dictionary need to be marked with the exact same conditions defined in the conditions
object. The isDictionaryForm
field can be set to false
, to allow some conditions to be used only in between rules, and not in the dictionary. In most cases however, it will be set to true
.
Now consider the word dogs'
, as in the the dogs' bones
. This is the possessive of a plural noun. We can add a rule for the possessive:
{
name: "possessive",
description: "Possessive form of a noun",
rules: [
suffixInflection("'", "", [], ["n"])
],
},
However, the only conditionOut
of this rule, n
, does not match any conditionIn
of the plural
rule, because the plural
rules conditionsIn
are an empty array. To fix this, we can add a condition to the plural
rule:
{
name: "plural",
description: "Plural form of a noun",
rules: [
suffixInflection("s", "", ["n"], ["n"])
],
},
Now the rules will chain together, as shown in the image. Chaining can be very useful (for agglutinative languages it is indispensable), but may cause unexpected behavior. For example, boss
will now display results for the word bo
(e.g. the staff) with the plural
rule applied twice, i.e. it can chain with itself because the conditionsIn
and conditionsOut
are the same. This leads us to the actual implementation of the plural
rule in english-transforms.js
:
conditions: {
n: {
name: "Noun",
isDictionaryForm: true,
subConditions: ["np", "ns"],
},
np: {
name: "Noun plural",
isDictionaryForm: true,
},
ns: {
name: "Noun singular",
isDictionaryForm: true,
},
},
transforms: {
"plural": {
name: "plural",
description: "Plural form of a noun",
rules: [
suffixInflection("s", "", ["np"], ["ns"])
],
},
},
Since ns
and np
are subconditions of n
they will both match with n
, but not with each other. This covers all of the requirements we have considered.
The suffixInflection
is one of a few helper functions - you can write more complex rules, using regex and a function for deinflecting. There are examples of this across the language transforms files.
Now that you have added a couple deinflection rules, you might want to start writing some tests to check if the deinflections are behaving correctly. Let's say we wanted to test the behavior of our plural
and possessive
rules and even them combined. Our test file should look like this:
// english-transforms.test.js
import { englishTransforms } from "../../ext/js/language/en/english-transforms.js";
import { LanguageTransformer } from "../../ext/js/language/language-transformer.js";
import { testLanguageTransformer } from "../fixtures/language-transformer-test.js";
const tests = [
{
category: "plurals and possessive",
valid: true,
tests: [
{ term: "cat", source: "cats", rule: "ns", reasons: ["plural"] },
{ term: "cat", source: "cat's", rule: "ns", reasons: ["possessive"] },
{
term: "cat",
source: "cats'",
rule: "ns",
reasons: ["plural", "possessive"],
},
],
},
];
const languageTransformer = new LanguageTransformer();
languageTransformer.addDescriptor(englishTransforms);
testLanguageTransformer(languageTransformer, tests);
The part we want to examine is the test
array. The other things are common across all test files.
term
is the final form of the deinflected word.source
is the source word to be deinflected toterm
.rule
is the final condition ofterm
. Here, we usedns
becausecat
is a singular noun.reasons
represents the chain of deinflection rules needed to get fromsource
toterm
.
You can check that all the tests pass by running npm run test:unit
.
This command runs all Rikaitan unit test files. To only run a single test file, you can instead opt for
npx vitest <file-name>
.
Now, we may want to verify that boss
really does not deinflect to bo
. You can add to the tests
array:
{
category: 'invalid deinflections',
valid: false,
tests: [
{term: 'boss', source: 'bo', rule: 'ns', reasons: ['plural', 'plural']},
],
},
Here, by setting valid
to false
, we are telling the test function to fail this test case if only boss
deinflects to bo
with the ns
condition under a double plural
rule.
You can also optionally pass a preprocess
helper function to testLanguageTransformer
. Refer to the language transforms test files for its specific use case.
If you want additional type-checking and autocompletion when writing your deinflection rules, you can add them with just a few extra lines of code. Due to the limitations of TypeScript and JSDoc annotations, we will have to perform some type magic in our transformations file, but you don't need to understand what they mean in detail.
Your english-transforms.js
file should look like this:
// english-transforms.js
import { suffixInflection } from "../language-transforms.js";
/** @type {import('language-transformer').LanguageTransformDescriptor} */
export const englishTransforms = {
language: "en",
conditions: {
n: {
name: "Noun",
isDictionaryForm: true,
subConditions: ["np", "ns"],
},
np: {
name: "Noun plural",
isDictionaryForm: true,
},
ns: {
name: "Noun singular",
isDictionaryForm: true,
},
},
transforms: {
// omitted
},
};
To gain type-safety, we have to pass an additional TCondition
type parameter to LanguageTransformDescriptor
. (You can revisit its definition at the top of this section)
The passed type value should be the union type of all conditions in our transforms. To find this value, we first need to move the conditions
object outside of englishTransforms
and extract its type by adding a /** @typedef {keyof typeof conditions} Condition */
comment at the start of the file. Then, you just need to pass it to the LanguageTransformDescriptor
type declaration like so:
// english-transforms.js
import { suffixInflection } from "../language-transforms.js";
/** @typedef {keyof typeof conditions} Condition */
const conditions = {
n: {
name: "Noun",
isDictionaryForm: true,
subConditions: ["np", "ns"],
},
np: {
name: "Noun plural",
isDictionaryForm: true,
},
ns: {
name: "Noun singular",
isDictionaryForm: true,
},
};
/** @type {import('language-transformer').LanguageTransformDescriptor<Condition>} */
export const englishTransforms = {
language: "en",
conditions,
transforms: {
// omitted
},
};
Now you should be able to check for types whenever writing a deinflection rule.
In special cases, text may need to be modified after deinflection. These work exactly like text preprocessors, but are applied after deinflection. Currently, this is only used for Korean, where the Hangul text is disassembled into jamo during preprocessing, and so must be reassembled after deinflection.
Some features include checking whether a string is possibly a word in the language. For example, trying to look up, in an English dictionary, a word written with non-Latin characters (e.g. "日本語") will never yield any results. To prevent unnecessary lookups, an isTextLookupWorthy
function can be provided, otherwise all text will be looked up.
In certain languages, dictionary entries may contain readings as a key to read words, e.g. Kana for Japanese and Pinyin for Chinese. Sometimes, dictionaries may be inconsistent in how they store these readings, leading to the word entries often being split when looked up even though they share the same reading. In these cases, you can use a readingNormalizer
function to normalize the readings to a common format.
If you have any questions, please feel free to open a Discussion on Github, or find us on DJT.