DSL::Bulgarian

Introduction

This blog post proclaims and exemplifies the Raku package “DSL::Bulgarian” that facilitates the specification of computational workflows using natural language commands in Bulgarian.

Using the Domain Specific Languages (DSLs) executable code is generated for different programming languages: Julia, Python, R, Raku, Wolfram Language (WL).

Translation to other natural languages is also done: English, Korean, Russian, Spanish.

The workflow of creating the package “DSL::Bulgarian” is also utilized for making of the packages “DSL::Portuguese” and “DSL::Russian”.

Why and how it works?

The package “DSL::Bulgarian” reuses the English-based grammars of the packages “DSL::English::*”, [AAp1 ÷ AAp8]. This produces good parsers of Bulgarian DSL commands because:

  • Bulgarian and English are both analytic languages.
  • The English grammar structures of short (imperative or prescriptive) sentences resemble those of Bulgarian.

It is relatively easy to “just” replace English tokens with Bulgarian tokens by direct translation. Further work is required, though, in order to handle Bulgarian word-suffixes related to gender and plurality.

Additionally, in English often a word can function as a noun and verb, or adjective and noun, or adjective, noun, and verb. Hence, the English grammars have to be (slightly) refactored to indicate what part of speech tokens are used and the corresponding English tokens files have to be enhanced.

Document organization

The document organization is as follows:

  • The next five sections show examples of code generation for computational workflows of the most fundamental Machine Learning (ML) sub-cultures: Data Wrangling, Classification, Latent Semantic Analysis (LSA), Quantile Regression (QR), Recommendations.
  • The following section discusses the general software engineering strategy, and related design and implementation.
  • The last section gives implementation details that exemplify some of the steps in the software strategy section.
  • Multiple references are given that provide further descriptions, explanations, and examples of the discussed features.

Data query (wrangling) workflows

Translate Bulgarian data wrangling specifications to different natural and programming languages:

use DSL::English::DataQueryWorkflows;

my $command = '
зареди данните iris;
вземи елементите от 1 до 120;
филтрирай чрез Sepal.Width е по-голямо от 2.4 и Petal.Length е по-малко от 5.5; 
групирай с колоната Species;
покажи размерите
';
for <English Python::pandas Raku::Reshapers Spanish Russian> -> $t {
   say '=' x 60, "\n", $t, "\n", '-' x 60;
   say ToDataQueryWorkflowCode($command, $t, language => 'Bulgarian', format => 'code');
}

# ============================================================
# English
# ------------------------------------------------------------
# load the data table: "iris"
# take elements from 1 to 120
# filter with the predicate: ((Sepal.Width greater than 2.4) и (Petal.Length less than 5.5))
# group by the columns: Species
# show the count(s)
# ============================================================
# Python::pandas
# ------------------------------------------------------------
# obj = example_dataset('iris')
# obj = obj.iloc[1-1:120]
# obj = obj[((obj["Sepal.Width"]> 2.4) & (obj["Petal.Length"]< 5.5))]
# obj = obj.groupby(["Species"])
# print(obj.size())
# ============================================================
# Raku::Reshapers
# ------------------------------------------------------------
# my $obj = example-dataset('iris') ;
# $obj = $obj[ (1 - 1) ... (120 - 1 ) ] ;
# $obj = $obj.grep({ $_{"Sepal.Width"} > 2.4 and $_{"Petal.Length"} < 5.5 }).Array ;
# $obj = group-by($obj, "Species") ;
# say "counts: ", $obj>>.elems
# ============================================================
# Spanish
# ------------------------------------------------------------
# cargar la tabla: "iris"
# tomar los elementos de 1 a 120
# filtrar con la condicion: ((Sepal.Width más grande 2.4) y (Petal.Length menos 5.5))
# agrupar con columnas: "Species"
# mostrar recuentos
# ============================================================
# Russian
# ------------------------------------------------------------
# загрузить таблицу: "iris"
# взять элементы с 1 по 120
# фильтровать с предикатом: ((Sepal.Width больше 2.4) и (Petal.Length меньше 5.5))
# групировать с колонками: Species
# показать число


Classification workflows

Translate Bulgarian ML classification specifications to different natural languages and WL:

use DSL::English::ClassificationWorkflows;

my $command = '
използвай dfTitanic;
раздели данните с цепещо съотношение 0.82;
направи gradient boosted trees класификатор;
покажи TruePositiveRate и FalsePositiveRate;
';

for <English Russian WL::ClCon> -> $t {
    say '=' x 60, "\n", $t, "\n", '-' x 60;
    say ToClassificationWorkflowCode($command, $t, language => 'Bulgarian', format => 'code');
}

# ============================================================
# English
# ------------------------------------------------------------
# use the data: dfTitanic 
# split into training and testing data with the proportion 0.82 
# train classifier with method: gradient boosted trees
# ============================================================
# Russian
# ------------------------------------------------------------
# использовать данные: dfTitanic 
# разделить данные на пропорцию 0.82 
# обучить классификатор методом: gradient boosted trees
# ============================================================
# WL::ClCon
# ------------------------------------------------------------
# ClConUnit[ dfTitanic ] \[DoubleLongRightArrow]
# ClConSplitData[ 0.82 ] \[DoubleLongRightArrow]
# ClConMakeClassifier[ "GradientBoostedTrees" ] \[DoubleLongRightArrow]
# ClConClassifierMeasurements[ {"Recall", "FalsePositiveRate"} ] \[DoubleLongRightArrow] ClConEchoValue[]

Remark: Currently, code generation for ML classification workflows is implemented only for WL.


Latent Semantic Analysis

Translate Bulgarian LSA specifications to different natural and programming languages:

use DSL::English::LatentSemanticAnalysisWorkflows;

my $command = '
създай със textHamlet;
направи документ-термин матрица със автоматични стоп думи;
приложи LSI функциите IDF, TermFrequency, и Cosine;
извади 12 теми чрез NNMF и максимален брой стъпки 12;
покажи таблица  на темите с 12 термина;
покажи текущата лентова стойност
';

for <English Python::LSAMon R::LSAMon Russian> -> $t {
    say '=' x 60, "\n", $t, "\n", '-' x 60;
    say ToLatentSemanticAnalysisWorkflowCode($command, $t, language => 'Bulgarian', format => 'code');
}

#ERROR: Possible misspelling of 'термини' as 'термина'.
#ERROR: Possible misspelling of 'термини' as 'термина'.
#ERROR: Possible misspelling of 'термини' as 'термина'.
#ERROR: Possible misspelling of 'термини' as 'термина'.
# ============================================================
# English
# ------------------------------------------------------------
# create LSA object with the data: textHamlet
# make the document-term matrix with the parameters: use the stop words: NULL
# apply the latent semantic analysis (LSI) functions: global weight function : "IDF", local weight function : "None", normalizer function : "Cosine"
# extract 12 topics using the parameters: method : Non-Negative Matrix Factorization (NNMF), max number of steps : 12
# show topics table using the parameters: numberOfTerms = 12
# show the pipeline value
# ============================================================
# Python::LSAMon
# ------------------------------------------------------------
# LatentSemanticAnalyzer(textHamlet).make_document_term_matrix( stop_words = None).apply_term_weight_functions(global_weight_func = "IDF", local_weight_func = "None", normalizer_func = "Cosine").extract_topics(number_of_topics = 12, method = "NNMF", max_steps = 12).echo_topics_table(numberOfTerms = 12).echo_value()
# ============================================================
# R::LSAMon
# ------------------------------------------------------------
# LSAMonUnit(textHamlet) %>%
# LSAMonMakeDocumentTermMatrix( stopWords = NULL) %>%
# LSAMonApplyTermWeightFunctions(globalWeightFunction = "IDF", localWeightFunction = "None", normalizerFunction = "Cosine") %>%
# LSAMonExtractTopics( numberOfTopics = 12, method = "NNMF",  maxSteps = 12) %>%
# LSAMonEchoTopicsTable(numberOfTerms = 12) %>%
# LSAMonEchoValue()
# ============================================================
# Russian
# ------------------------------------------------------------
# создать латентный семантический анализатор с данных: textHamlet
# сделать матрицу документов-терминов с параметрами: стоп-слова: null
# применять функции латентного семантического индексирования (LSI): глобальная весовая функция: "IDF", локальная весовая функция: "None", нормализующая функция: "Cosine"
# извлечь 12 тем с параметрами: метод: Разложение Неотрицательных Матричных Факторов (NNMF), максимальное число шагов: 12
# показать таблицу темы по параметрам: numberOfTerms = 12
# показать текущее значение конвейера

Remark: The misspelling prompts show that the misspellings handler is “too aggressive.” The module of handling DSL mis-formulations can be improved with several different approaches. Currently, only token-level fuzzy matching is used. Other approaches are: (i) using an NLP Template Engine, [AAr1], (ii) phrase- or command-level fuzzy matching, (iii) conversational agents dialogs, [AAv1].


Quantile Regression Workflows

Translate Bulgarian QR specifications to different natural and programming languages:

use DSL::English::QuantileRegressionWorkflows;

my $command = '
създай с dfTemperatureData;
премахни липсващите стойности;
покажи данново обобщение;
премащабирай двете оси;
изчисли квантилна регресия с 20 възела и вероятности от 0.1 до 0.9 със стъпка 0.1;
покажи диаграма с дати;
покажи чертеж на абсолютните грешки;
покажи текущата лентова стойност
';

for <English R::QRMon Russian WL::QRMon> -> $t {
    say '=' x 60, "\n", $t, "\n", '-' x 60;
    say ToQuantileRegressionWorkflowCode($command, $t, language => 'Bulgarian', format => 'code');
}

#ERROR: Possible misspelling of 'възли' as 'възела'.
#ERROR: Possible misspelling of 'възли' as 'възела'.
#ERROR: Possible misspelling of 'възли' as 'възела'.
#ERROR: Possible misspelling of 'възли' as 'възела'.
# ============================================================
# English
# ------------------------------------------------------------
# create quantile regression object with the data: dfTemperatureData
# delete missing values
# show data summary
# rescale: over both regressor and value axes
# compute quantile regression with parameters: degrees of freedom (knots): 20, automatic probabilities
# show plot with parameters: use date axis
# show plot of relative errors
# show the pipeline value
# ============================================================
# R::QRMon
# ------------------------------------------------------------
# QRMonUnit( data = dfTemperatureData) %>%
# QRMonDeleteMissing() %>%
# QRMonEchoDataSummary() %>%
# QRMonRescale(regressorAxisQ = TRUE, valueAxisQ = TRUE) %>%
# QRMonQuantileRegression(df = 20, probabilities = seq(0.1, 0.9, 0.1)) %>%
# QRMonPlot( datePlotQ = TRUE) %>%
# QRMonErrorsPlot( relativeErrorsQ = TRUE) %>%
# QRMonEchoValue()
# ============================================================
# Russian
# ------------------------------------------------------------
# создать объект квантильной регрессии с данными: dfTemperatureData
# удалить пропущенные значения
# показать сводку данных
# перемасштабировать: по осям регрессии и значений
# рассчитать квантильную регрессию с параметрами: степени свободы (узлы): 20, автоматическими вероятностями
# показать диаграмму с параметрами: использованием оси дат
# показать диаграму на относительных ошибок
# показать текущее значение конвейера
# ============================================================
# WL::QRMon
# ------------------------------------------------------------
# QRMonUnit[dfTemperatureData] \[DoubleLongRightArrow]
# QRMonDeleteMissing[] \[DoubleLongRightArrow]
# QRMonEchoDataSummary[] \[DoubleLongRightArrow]
# QRMonRescale["Axes"->{True, True}] \[DoubleLongRightArrow]
# QRMonQuantileRegression["Knots" -> 20, "Probabilities" -> Range[0.1, 0.9, 0.1]] \[DoubleLongRightArrow]
# QRMonDateListPlot[] \[DoubleLongRightArrow]
# QRMonErrorPlots[ "RelativeErrors" -> True] \[DoubleLongRightArrow]
# QRMonEchoValue[]


Recommender workflows

Translate Bulgarian ML recommendation specifications to different natural and programming languages:

use DSL::English::RecommenderWorkflows;

my $command = '
създай чрез dfTitanic;
препоръчай със профила "male" и "died";
покажи текущата лентова стойност
';

for <English Python::SMRMon R::SMRMon Russian> -> $t {
    say '=' x 60, "\n", $t, "\n", '-' x 60;
    say ToRecommenderWorkflowCode($command, $t, language => 'Bulgarian', format => 'code');
}

# ============================================================
# English
# ------------------------------------------------------------
# create with data table: dfTitanic
# recommend with the profile: ["male", "died"]
# show the pipeline value
# ============================================================
# Python::SMRMon
# ------------------------------------------------------------
# obj = SparseMatrixRecommender().create_from_wide_form(data = dfTitanic).recommend_by_profile( profile = ["male", "died"]).echo_value()
# ============================================================
# R::SMRMon
# ------------------------------------------------------------
# SMRMonCreate(data = dfTitanic) %>%
# SMRMonRecommendByProfile( profile = c("male", "died")) %>%
# SMRMonEchoValue()
# ============================================================
# Russian
# ------------------------------------------------------------
# создать с таблицу: dfTitanic
# рекомендуй с профилю: ["male", "died"]
# показать текущее значение конвейера


Software development strategy

General steps

In order to have Bulgarian commands parsed and interpreted into code the steps taken were split into four phases:

  1. Utilities preparation
  2. Bulgarian words and phrases addition and preparation
  3. Preliminary functionality experiments
  4. Packages code refactoring

The big picture

The following flowchart summarizes the total Raku for Prediction system development and its extension to with Bulgarian DSL code generation.

The sub-sections below give a corresponding (coarsely grained) narration.

Tool and targets

Here are the translations tool and targets:

  • Implementation tool is Raku.
  • Primary (translation) targets are Mathematica / WL and R.
  • Secondary targets are Python, Bulgarian, English, Russian.
  • Tertiary targets are Julia, Korean, Spanish.

Ground work

Here are the ground work steps of he Raku for Prediction system:

  1. First extract, define, and implement into software monads the computational workflows for
    Data Wrangling, Classification, Latent Semantic Analysis, Quantile Regression, Recommendations.
  2. Implement code transformation packages that enhance Raku grammar and role files with additional specifications.
    • For example, tokens and rules are modified to do fuzzy matching (using Levenshtein distance.)
  3. Make a “solid” implementation of Data Wrangling DSL translations from English into Julia, Python, R, Raku, and WL; see [AAp5].
  4. Make a comprehensive translator with a Web service.
  5. Implement multi-language numeric word forms parsers. For example:
use Lingua::NumericWordForms;
from-numeric-word-form(['twenty five', 'двадесет и пет', 'двадцать пять']):p

# (english => 25 bulgarian => 25 russian => 25)

  1. Implement Raku-grammar generation packages from Java’s ANTLR and WL.
    • In order to leverage “prior art” of DSL translations.
  2. Implement entity name parsers. For example:
use DSL::Entity::MachineLearning;
ToMachineLearningEntityCode('true positive rate')
# "Recall"
  1. Implement “higher order” DSL translations for problem domains that might utilize ML workflows.
    • Like data acquisition, [AAp9], food preparation, [AAp10], or recruiting, [AAp11]. See [AAv1] for more detailed discussion.
  2. Implement Raku packages for data wrangling; see [AAv2].
  3. Implement alternative translation approach using neural networks — instead of grammars; see [AAr1, AAv4]. For example, here the neural-network-based NLP template engine is utilized:
dsl-web-translation "make some sort of recommender pipeline over the dataset dfGoods"

# {"CODE":"smrObj <-\nSMRMonUnit() %>%\nSMRMonCreate( data = dfGoods) %>%\nSMRMonRecommendByProfile( profile = c(\"recommender\"), nrecs = 12) %>%\nSMRMonJoinAcross( data = dfGoods) %>%\nSMRMonEchoValue()","USERID":"","DSLTARGET":"R::Recommendations","DSL":"Recommendations","DSLFUNCTION":"Concretize[Recommendations, #1, TargetLanguage -> R, AvoidMonads -> False, AssociationResult -> True] & "}

  1. Design and implement Finite State Machine (FSM) framework in order to facilitate the implementation of conversational agents with “state remembering dialogs.” See [AAv1].

Bulgarian DSL interpreters implementation

  1. Implement Bulgarian fuzzy matching. See is-bg-fuzzy-match in “DSL::Shared”, [AAp2].
  2. Productize the Command Line Interface (CLI) scripts of “Grammar::TokenProcessing”, [AAp3], to a full extent.
  3. Refactor English phrases template files to clearly indicate DSL tokens type (adjectives, nouns, verbs.)
  4. Refactor grammars and code generators.
    • Introduce the Grammarish roles into each of the focus DSL packages (“DSL::English::*”.)
  5. Make translator actions for Bulgarian, English, and Russian.
    • This means, for example, that “DSL::English::ClassificationWorkflows” has actions:
      • “DSL::English::ClassificationWorkflows::Actions::Bulgarian::Standard”
      • “DSL::English::ClassificationWorkflows::Actions::English::Standard”
      • “DSL::English::ClassificationWorkflows::Actions::Russian::Standard”
  6. Make the English to Bulgarian “direct” translation of tokens. (As described in the introduction.)
  7. Create and run unit tests.
  8. Enhance entity packages to utilize Bulgarian names of entities.
  9. Make the higher order DSLs work with Bulgarian. (Not done yet.)
  10. Write documentation and make presentations. (Started; see [AA2].)
  11. Hook up the Bulgarian parser-interpreters to the Web service mentioned above. For example:
dsl-web-translation -t=WL "DSL MODULE ClCon;
използвай dfTitanic; 
раздели данните с пропорция 0.8; 
направи класификатор с метода RandomForest; 
покажи мерките"

# ClConUnit[ dfTitanic ] \[DoubleLongRightArrow]
# ClConSplitData[ "TrainingFraction" -> 0.8 ] \[DoubleLongRightArrow]
# ClConMakeClassifier[ "RandomForest" ] \[DoubleLongRightArrow]
# ClConClassifierMeasurements[ {"Precision", "Recall"} ] \[DoubleLongRightArrow] ClConEchoValue[]


Implementation details

Utilities preparation

Since the beginning of the work on translation of the computational DSLs into programming code it was clear that some the required code transformations have to be automated.

While doing the preparation work — and in general, while the DSL-translation work matured — it became clear that there are several directives to follow:

  1. Make and use Command Line Interface (CLI) scripts that do code transformation or generation.
  2. Adhere to of the Eric Raymond’s 17 Unix Rules, [Wk1]:
    • Make data complicated when required, not the program
    • Write abstract programs that generate code instead of writing code by hand

In order to facilitate the “from Bulgarian” project the package “Grammar::TokenProcessing”, [AAp3], was “finalized.” The initial versions of that package were used from the very beginning of the DSLs grammar development in order to facilitate handling of misspellings.

Current recipe

This sub-section lists the steps for endowing a certain already developed workflows DSL package with Bulgarian translations.

Denote the DSL workflows we focus on as DOMAIN (workflows.) For example, DOMAIN can stand for DataQueryWorkflows, or RecommenderWorkflows.

Remark: In the recipe steps below DOMAIN would be DataQueryWorkflows

It is assumed that:

  • DOMAIN in English are already developed.
  • Since both English and Bulgarian are analytical, non-agglutinative languages “just” replacing English words with Bulgarian words in DOMAIN would produce good enough parsers of Bulgarian.

Here are the steps:

  1. Add global Bulgarian words (optional)
    1. Add Bulgarian words and phrases in the DSL::Shared file “Roles/Bulgarian/CommonSpeechParts-template”.
    2. Generate the file Roles/Bulgarian/CommonSpeechParts.rakumod using the CLI script AddFuzzyMatching
    3. Consider translating, changing, or refactoring global files, like, Roles/English/TimeIntervalSpec.rakumod
  2. Translate DOMAIN English words and phrases into Bulgarian
    1. Take the file DOMAIN/Grammar/DOMAIN-template and translate its words into Bulgarian
  3. Add the corresponding files into DSL::Bulgarian, [AAp1].
    1. Use the DOMAIN/Grammarish.rakumod role.
      • The English DOMAIN package should have such rule. If do not do the corresponding code refactoring.
    2. Test with implemented DOMAIN languages.
    3. See the example grammar and role in DataQueryWorkflows in DSL::Bulgarian.

Code modification examples

The rules in the file “DataQueryPhrases.rakumod” are derived from file “DataQueryPhrases-template” using the package “Grammar::TokenProcessing” , [AAp3].


References

Articles

[AA1] Anton Antonov, “Introduction to data wrangling with Raku”, (2021), RakuForPrediction at WordPress.

[AA2] Anton Antonov, “Увод в обработката на данни с Raku”, (2022), RakuForPrediction at WordPress.

[Wk1] Wikipedia entry, UNIX-philosophy rules.

Packages

[AAp1] Anton Antonov, DSL::Bulgarian, Raku package, (2022), GitHub/antononcube.

[AAp2] Anton Antonov, DSL::Shared, Raku package, (2018-2022), GitHub/antononcube.

[AAp3] Anton Antonov, Grammar::TokenProcessing, Raku project (2022), GitHub/antononcube.

[AAp4] Anton Antonov, DSL::English::ClassificationWorkflows, Raku package, (2018-2022), GitHub/antononcube.

[AAp5] Anton Antonov, DSL::English::DataQueryWorkflows, Raku package, (2020-2022), GitHub/antononcube.

[AAp6] Anton Antonov, DSL::English::LatentSemanticAnalysisWorkflows, Raku package, (2018-2022), GitHub/antononcube.

[AAp7] Anton Antonov, DSL::English::QuantileRegressionWorkflows, Raku package, (2018-2022), GitHub/antononcube.

[AAp8] Anton Antonov, DSL::English::RecommenderWorkflows, Raku package, (2018-2022), GitHub/antononcube.

[AAp9] Anton Antonov, DSL::English::DataAcquisitionWorkflows, Raku package, (2021-2022), GitHub/antononcube.

[AAp10] Anton Antonov, DSL::English::FoodPreparationWorkflows, Raku package, (2021-2022), GitHub/antononcube.

[AAp11] Anton Antonov, DSL::English::RecruitingWorkflows, Raku package, (2021-2022), GitHub/antononcube.

Repositories

[AAr1] Anton Antonov, “NLP Template Engine”, (2021-2022), GitHub/antononcube.

Videos

[AAv1] Anton Antonov, “Multi-language Data Wrangling and Acquisition Conversational Agents”, (2022), FOSDEM 2022.

[AAv2] Anton Antonov, “Implementing Machine Learning algorithms in Raku”, (2022), The Raku Conference 2022.

[AAv3] Anton Antonov, “Simplified Machine Learning Workflows Overview”, (2022), Wolfram Technology Conference 2022.

[AAv4] Anton Antonov, “Natural Language Processing Template Engine”, (2022), Wolfram Technology Conference 2022.

Advertisement

2 thoughts on “DSL::Bulgarian

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s