Advertisement
Guest User

Pseudocode

a guest
Mar 16th, 2020
194
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.60 KB | None | 0 0
  1. let english_words be "the most frequent 20,000 words taken from all available manuscripts on google books from 1600 to 1850"
  2.  
  3. for each language in all_languages:
  4. let phonemes[language] be "a list of phonemes and associated ligatures for language, generated according to some hidden parameters based on the language"
  5. initialize "generatedWords[language]" collection
  6. for each word in english_words:
  7. let expected_length be "length of word in english divided by number of letters in english times number of letters in language"
  8. if word shares a stem with an already generated word:
  9. newWord = combine(language, shared_stem, generate_word(word, language, expected_length - stem_length))
  10. else:
  11. newWord = generate_word(word, language, expected_length)
  12. generatedWords[language].Add(newWord)
  13.  
  14. def combine:
  15. get language rule for prefix, infix, postfix
  16. combine stem with partial word according to rule
  17. return result
  18.  
  19. def generate_word:
  20. generated_word = ""
  21. for each phoneme in word:
  22. select a phoneme from phonemes[language] with approximately similar frequency to the word's phoneme in english
  23. generated_word += ligature associated with selected phoneme
  24. apply some rules for truncation/contraction according to hidden language parameters
  25. apply some prefix/suffix rules according to hidden language parameters
  26. if length of generated_word is greater than expected_length or word has too many of the same symbols:
  27. regenerate the word with probability proportional to misfit
  28. return generated_word
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement