diff --git a/wikispeech_server/adapters/test_lexicon_client.py b/wikispeech_server/adapters/test_lexicon_client.py index 34f826e..1abff37 100644 --- a/wikispeech_server/adapters/test_lexicon_client.py +++ b/wikispeech_server/adapters/test_lexicon_client.py @@ -16,12 +16,12 @@ class TestLexicon(unittest.TestCase): def testNewLexicon(self): - lexicon_name = "pronlex:sv-se.nst" + lexicon_name = "sv_se_nst_lex:sv-se.nst" lexicon = Lexicon(lexicon_name) self.assertEqual(str(type(lexicon)), "") def testLookup(self): - lexicon_name = "pronlex:sv-se.nst" + lexicon_name = "sv_se_nst_lex:sv-se.nst" lexicon = Lexicon(lexicon_name) orth = "apa" @@ -30,7 +30,7 @@ def testLookup(self): #expected = [{'entryValidations': [], 'partOfSpeech': 'NN', 'language': 'sv-se', 'transcriptions': [{'id': 79410, 'entryId': 74074, 'sources': [], 'language': 'sv-se', 'strn': '"" A: . p a'}], 'id': 74074, 'preferred': False, 'morphology': 'SIN|IND|NOM|UTR', 'lemma': {'id': 8764, 'paradigm': 's1a-flicka', 'reading': '', 'strn': 'apa'}, 'wordParts': 'apa', 'strn': 'apa', 'lexiconId': 1, 'status': {'name': 'imported', 'id': 74074, 'current': True, 'source': 'nst', 'timestamp': '2017-05-12T10:55:49Z'}}, {'entryValidations': [], 'partOfSpeech': 'VB', 'language': 'sv-se', 'transcriptions': [{'id': 79411, 'entryId': 74075, 'sources': [], 'language': 'sv-se', 'strn': '"" A: . p a'}], 'id': 74075, 'preferred': False, 'morphology': '', 'lemma': {'id': 8764, 'paradigm': 's1a-flicka', 'reading': '', 'strn': 'apa'}, 'wordParts': 'apa', 'strn': 'apa', 'lexiconId': 1, 'status': {'name': 'imported', 'id': 74075, 'current': True, 'source': 'nst', 'timestamp': '2017-05-12T10:55:49Z'}}, {'entryValidations': [], 'partOfSpeech': 'VB', 'language': 'sv-se', 'transcriptions': [{'id': 79412, 'entryId': 74076, 'sources': [], 'language': 'sv-se', 'strn': '"" A: . p a'}], 'id': 74076, 'preferred': False, 'morphology': 'AKT|INF-IMP', 'lemma': {'id': 8764, 'paradigm': 's1a-flicka', 'reading': '', 'strn': 'apa'}, 'wordParts': 'apa', 'strn': 'apa', 'lexiconId': 1, 'status': {'name': 'imported', 'id': 74076, 'current': True, 'source': 'nst', 'timestamp': '2017-05-12T10:55:49Z'}}] - expected = [{'lexRef': {'DBRef': 'pronlex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74074, 'strn': '"" A: . p a', 'id': 79410, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'NN', 'wordParts': 'apa', 'id': 74074, 'morphology': 'SIN|IND|NOM|UTR', 'status': {'source': 'nst', 'id': 74074, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}, {'lexRef': {'DBRef': 'pronlex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74075, 'strn': '"" A: . p a', 'id': 79411, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'VB', 'wordParts': 'apa', 'id': 74075, 'morphology': '', 'status': {'source': 'nst', 'id': 74075, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}, {'lexRef': {'DBRef': 'pronlex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74076, 'strn': '"" A: . p a', 'id': 79412, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'VB', 'wordParts': 'apa', 'id': 74076, 'morphology': 'AKT|INF-IMP', 'status': {'source': 'nst', 'id': 74076, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}] + expected = [{'lexRef': {'DBRef': 'sv_se_nst_lex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74074, 'strn': '"" A: . p a', 'id': 79410, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'NN', 'wordParts': 'apa', 'id': 74074, 'morphology': 'SIN|IND|NOM|UTR', 'status': {'source': 'nst', 'id': 74074, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}, {'lexRef': {'DBRef': 'sv_se_nst_lex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74075, 'strn': '"" A: . p a', 'id': 79411, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'VB', 'wordParts': 'apa', 'id': 74075, 'morphology': '', 'status': {'source': 'nst', 'id': 74075, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}, {'lexRef': {'DBRef': 'sv_se_nst_lex', 'LexName': 'sv-se.nst'}, 'lemma': {'strn': 'apa', 'paradigm': 's1a-flicka', 'reading': '', 'id': 8764}, 'strn': 'apa', 'transcriptions': [{'entryId': 74076, 'strn': '"" A: . p a', 'id': 79412, 'language': 'sv-se', 'sources': []}], 'preferred': False, 'partOfSpeech': 'VB', 'wordParts': 'apa', 'id': 74076, 'morphology': 'AKT|INF-IMP', 'status': {'source': 'nst', 'id': 74076, 'timestamp': '2017-08-17T11:57:08Z', 'name': 'imported', 'current': True}, 'language': 'sv-se', 'entryValidations': []}] result = lexicon.lookup(orth) log.info("RESULT: %s" % result) @@ -41,7 +41,7 @@ def testLookup(self): def testLexiconException1(self): default_log_level = log.log_level log.log_level = "fatal" - lexicon_name = "pronlex:sv-se.nst_THIS_LEXICON_SHOULD_NOT_EXIST" + lexicon_name = "sv_se_nst_lex:sv-se.nst_THIS_LEXICON_SHOULD_NOT_EXIST" with self.assertRaises(LexiconException): lexicon = Lexicon(lexicon_name) log.log_level = default_log_level @@ -49,9 +49,9 @@ def testLexiconException1(self): def testLexiconException2(self): default_log_level = log.log_level log.log_level = "fatal" - lexicon_name = "pronlex:sv-se.nst_THIS_LEXICON_SHOULD_NOT_EXIST" + lexicon_name = "sv_se_nst_lex:sv-se.nst_THIS_LEXICON_SHOULD_NOT_EXIST" with self.assertRaises(LexiconException): - lexicon = Lexicon("pronlex:sv-se.nst") + lexicon = Lexicon("sv_se_nst_lex:sv-se.nst") lexicon.lexicon_name = lexicon_name lexicon.lookup("apa") log.log_level = default_log_level @@ -61,7 +61,7 @@ def test_lexLookup(self): lex_config = { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst" + "lexicon":"sv_se_nst_lex:sv-se.nst" } utt = { "lang": "sv", @@ -99,7 +99,7 @@ def test_lexLookup_Exception(self): lex_config = { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst_DOES_NOT_EXIST" + "lexicon":"sv_se_nst_lex:sv-se.nst_DOES_NOT_EXIST" } utt = { "lang": "sv", diff --git a/wikispeech_server/test_textprocessor.py b/wikispeech_server/test_textprocessor.py index 20afaf0..ce4852b 100644 --- a/wikispeech_server/test_textprocessor.py +++ b/wikispeech_server/test_textprocessor.py @@ -30,7 +30,7 @@ def testNewTextprocessor(self): { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst" + "lexicon":"sv_se_nst_lex:sv-se.nst" } ] } @@ -57,7 +57,7 @@ def testBrokenTextprocessor(self): { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst" + "lexicon":"sv_se_nst_lex:sv-se.nst" } ] } diff --git a/wikispeech_server/textprocessor.py b/wikispeech_server/textprocessor.py index a2d28bc..9accd1d 100644 --- a/wikispeech_server/textprocessor.py +++ b/wikispeech_server/textprocessor.py @@ -101,7 +101,7 @@ def __init__(self, cconfig): { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst" + "lexicon":"sv_se_nst_lex:sv-se.nst" } ] } diff --git a/wikispeech_server/voice_config.py b/wikispeech_server/voice_config.py index de83bd3..c088256 100644 --- a/wikispeech_server/voice_config.py +++ b/wikispeech_server/voice_config.py @@ -14,7 +14,7 @@ { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:sv-se.nst" + "lexicon":"sv_se_nst_lex:sv-se.nst" } ] } @@ -49,7 +49,7 @@ { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:en-us.cmu" + "lexicon":"en_am_cmu_lex:en-us.cmu" } ] } @@ -66,7 +66,7 @@ { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:en-us.cmu" + "lexicon":"en_am_cmu_lex:en-us.cmu" } ] } @@ -97,7 +97,7 @@ { "module":"adapters.lexicon_client", "call":"lexLookup", - "lexicon":"pronlex:ar-test" + "lexicon":"ar_ar_tst_lex:ar-test" } ] } diff --git a/wikispeech_server/wikispeech.py b/wikispeech_server/wikispeech.py index a063d0d..e571105 100644 --- a/wikispeech_server/wikispeech.py +++ b/wikispeech_server/wikispeech.py @@ -753,7 +753,7 @@ def getParam(param,default=None): def test_lexicon_client(): - lexicon = "pronlex:sv-se.nst" + lexicon = "sv_se_nst_lex:sv-se.nst" sent = "apa hund färöarna" trans = {} trans["apa"] = '"" A: . p a' diff --git a/workflow_demo/ko_test.html b/workflow_demo/ko_test.html index 30b3eaa..3c8b6e0 100644 --- a/workflow_demo/ko_test.html +++ b/workflow_demo/ko_test.html @@ -71,7 +71,7 @@ self.entries([]); var params = { - "lexicons": "pronlex:sv-se.nst", + "lexicons": "sv_se_nst_lex:sv-se.nst", "words": search_term } diff --git a/workflow_demo/test_workflow.js b/workflow_demo/test_workflow.js index db08de2..bd72a0b 100644 --- a/workflow_demo/test_workflow.js +++ b/workflow_demo/test_workflow.js @@ -884,10 +884,10 @@ function searchLexicon(search_term, lang) { if ( lang == "sv" ) { - var lexicons = "pronlex:sv-se.nst"; + var lexicons = "sv_se_nst_lex:sv-se.nst"; } else if ( lang == "en" ) { - var lexicons = "pronlex:en-us.cmu"; + var lexicons = "en_am_cmu_lex:en-us.cmu"; } else { console.log("WARNING: no lexicon defined for lang "+lang); @@ -917,7 +917,7 @@ function wordInLex(word, div, trans) { var unknown_words_container = document.getElementById("unknown_words_container"); var params = { - "lexicons": "pronlex:sv-se.nst", + "lexicons": "sv_se_nst_lex:sv-se.nst", "words": word } @@ -955,10 +955,10 @@ function wordsInLex(words, lang) { //TODO hardcoded lexicon if ( lang == "sv" ) { - var lexicons = "pronlex:sv-se.nst"; + var lexicons = "sv_se_nst_lex:sv-se.nst"; } else if ( lang == "en" ) { - var lexicons = "pronlex:en-us.cmu"; + var lexicons = "en_am_cmu_lex:en-us.cmu"; } else { console.log("WARNING: no lexicon defined for lang "+lang);