Wikifunctions is a new site that has been added to the list of sites operated by WMF. I definitely see uses for it in automating updates on Wikipedia and bots (and also for programmers to reference), but their goal is to translate Wikipedia articles to more languages by writing them in code that has a lot of linguistic information. I have mixed feelings about this, as I don’t like existing programs that automatically generate articles (see the Cebuano and Dutch Wikipedias), and I worry that the system will be too complicated for average people.
Sounds like a great idea. Plain English (or any human language) is not the best way to store information. I’ve certainly noticed mismatches between the data in different languages, or across related articles, because they don’t share the same data source.
Take a look at the article for NYC in English and French and you’ll see a bunch of data points, like total area, that are different. Not huge differences, but any difference at all is enough to demonstrate the problem. There should be one canonical source of data shared by all representations.
Wikipedia is available in hundreds of languages. Why should hundreds of editors need to update the NYC page every time a new census comes out with new population numbers? Ideally, that would require only one change to update every version of the article.
In programming, the convention is to separate the data from the presentation. In this context, plain-English is the presentation, and weaving actual data into it is sub-optimal. Something like population or area size of a city is not language-dependent, and should not be stored in a language-dependent way.
Ultimately, this is about reducing duplicate effort and maintaining data integrity.
This problem was solved in like 2012 or 2013 with the introduction of Wikidata, but not all language editions have decided to use that.
How common is it in English? I haven’t checked a lot of articles, but I did check the source of the English and French NYC articles I linked and it seems like all the information is hardcoded, not referenced from Wikidata.
I think enwiki tends to use Wikidata relatively sparingly.
Removed by mod
Disclaimer, I didn’t do any research on this, but what would be bad with just AI translating text, given a reliable enough AI? No code required, just plain human speech.
This will help make machine translation more reliable, ensuring that objective data does not get transformed along with the language presenting that data. It will also make it easier to test and validate the machine translators.
Any automated translations would still need to reviewed. I don’t think we will (or should) see totally automated translations in the near future, but I do think the machine translators could be a very useful tool for editors.
Language models are impressive, but they are not efficient data retrieval systems. Denny Vrandecic, the founder of Wikidata, has a couple insightful videos about this topic.
This one talks about knowledge graphs in general, from 2020: https://www.youtube.com/watch?v=Oips1aW738Q
This one is from last year and is specifically about how you could integrate LLMs with the knowledge graph to greatly increase their accuracy, utility, and efficiency: https://www.youtube.com/watch?v=WqYBx2gB6vA
I highly recommend that second video. He does a great job laying out what LLMs are efficient for, what more conventional methods are efficient for, and how you can integrate them to get the best of both worlds.
Thanks! I’ll come back to this thread once I read more.
Here is an alternative Piped link(s):
https://www.piped.video/watch?v=Oips1aW738Q
https://www.piped.video/watch?v=WqYBx2gB6vA
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I’m open-source; check me out at GitHub.
I assume the main benefit will be for users of less-spoken languages, who currently get out-of date articles or none at all.
but their goal is to translate Wikipedia articles to more languages by writing them in code that has a lot of linguistic information
That’ll get unruly really fast.
Languages simply don’t agree on how to split the usage of words. Or grammatical case. Or if, when and how to do agreement.
Just for the sake of example: how are they going to keep track of case in a way that doesn’t break Hindi, or Basque, or English, or Guarani? Or grammatical gender for a word like “milk”? (not even the Romance languages agree in it.) At a certain point, it gets simply easier to write the article in all those languages than to code something to make it for you.
I think that the best use scenario is to automate tidbits of highly changing data. It’s fairly limited but it could be useful.
Languages simply don’t agree on how to split the usage of words. Or grammatical case. Or if, when and how to do agreement.
Just for the sake of example: how are they going to keep track of case in a way that doesn’t break Hindi, or Basque, or English, or Guarani? Or grammatical gender for a word like “milk”? (not even the Romance languages agree in it.) At a certain point, it gets simply easier to write the article in all those languages than to code something to make it for you.
I don’t know what the WMF is planning here but what you’re pointing out is precisely what abstraction would solve.
If you had an abstract way to represent a sentence, you would be independent of any one order or case or whatever other grammatical feature. In the end you obviously do need actual sentences with these features. To get these, you’d build a mechanism that would convert the abstract sentence representation into a concrete sentences for specific languages that is correctly constructed according to those specific languages’ rules.
Same with gender. What you’d store would not be that e.g. some german sentence is talking about the feminine milk but rather that it’s talking about the abstract concept of milk. How exactly that abstract concept is represented in words would then be up to individual languages to decide.
I have absolutely no idea whether what I’m talking about here would be practical to implement but it in theory it could work.
Abstractions are not magic, and they cannot make info appear out of nowhere. Somewhere inside that abstraction you’ll need to have the pieces of info that Spanish “leche” [milk] is feminine, that Zulu “ubisi” [milk] is class 11, that English predicative uses the ACC form, so goes on.
And you’ll need people to mark a multitude of distinctions in their sentences, when writing them down, that the abstraction layer would demand for other languages. Such as tagging the “I” in “I see a boy” as “+masculine, +older-person, +informal” so Japanese correctly conveys it as “ore” instead of “boku”, "atashi, “watashi” etc.
Even the idea of “abstract concept of milk” doesn’t work as well as it sounds like, because languages will split even the abstract concepts in different ways. For example, does the abstract concept associated with a living pig includes its flesh?
And the language itself cannot decide those things. A language is not an agent; it doesn’t “do” something. You’d need people to actively insert those pieces of info for each language, that’s perhaps doable for the most spoken ones, but those are the ones that would benefit the least from this.
Somewhere inside that abstraction you’ll need to have the pieces of info that Spanish “leche” [milk] is feminine, that Zulu “ubisi” [milk] is class 11, that English predicative uses the ACC form, so goes on.
Of course you do. The beauty of abstraction is that these language-specific parts can be factored into generic language-specific components. The information you’re actually trying to convey can be denoted without any language-specific parts or exceptions and that’s the important part for Wikipedia’s purpose of knowledge preservation and presentation.
you’ll need people to mark a multitude of distinctions in their sentences, when writing them down, that the abstraction layer would demand for other languages. Such as tagging the “I” in “I see a boy” as “+masculine, +older-person, +informal” so Japanese correctly conveys it as “ore” instead of “boku”, "atashi, “watashi” etc.
For writing a story or prose, I agree.
For the purpose of writing Wikipedia articles, this specifically and explicitly does not matter very much. Wikipedia strives to have one unified way of writing within a language. Whether the “I” is masculine or not would be a parameter that would be applied to all text equally (assuming I-narrator was the standard on Wikipedia).
Even the idea of “abstract concept of milk” doesn’t work as well as it sounds like, because languages will split even the abstract concepts in different ways. For example, does the abstract concept associated with a living pig includes its flesh?
If your article talks about the concept of a living pig in some way and in the context of that article, it doesn’t matter whether the flesh is included, then you simply use the default word/phrase that the language uses to convey the concept of a pig.
If it did matter, you’d explicitly describe the concept of “a living pig with its flesh” instead of the more generic concept of a living pig. If that happened to be the default of the target language or the target language didn’t differentiate between the two concepts, both concepts would turn into the same terms in that specific language.
The same applies to your example of the different forms of “I” in Japanese. To create an appropriate Japanese “rendering” of an abstract sentence, you’d use the abstract concept of “a nerdy shy kid refers to itself” as the i.e. the subject. The Japanese language “renderer” would turn that into a sentence like ”僕は。。。” while the English “renderer” would simply produce “I …”.
A language is not an agent; it doesn’t “do” something. You’d need people to actively insert those pieces of info for each language, that’s perhaps doable for the most spoken ones, but those are the ones that would benefit the least from this.
Yes, of course they would have to do that. The cool thing is that this it’d only have to be done once in a generic manner and from that point on you could use that definition to “render” any abstract article into any language you like.
You must also keep in mind that this effort has to be measured relative to the alternatives. In this case, the alternative is to translate each and every article and all changes done to them into every available language. At the scale of Wikipedia, that is not an easy task and it’s been made clear that that’s simply not happening.
(Okay, another alternative would be to remain on the status quo with its divergent versions of what are supposed to be the same articles containing the same information.)
Note: I’ll clip the quotes for succinctness.
Of course you do. […]
You can’t leave those things to the abstraction layer because how different languages map abstract concepts differs, so there’s no way to factor them into generic language-specific components. The writer will need to tag things down, to minimal details, for the sake of languages that they don’t care about. It ends like that story about a map so large that it represents the terrain accurately being as big as the terrain, thus useless.
For writing a story or prose, I agree. […]
As I said in the reply to the other poster, the first pronoun is an example. This issue affects languages as a whole, and sometimes in ways that you can’t arbitrate through a fixed writing style because they convey meaning. (For example: if you don’t encode the social gender into the 3rd person pronouns, English breaks.)
If your article talks about the concept of a living pig in some way and in the context of that article, it doesn’t matter whether the flesh is included, then you simply use the default word/phrase that the language uses to convey the concept of a pig. […]
Often there’s no such thing as the “default”. The example with pig/pork is one of those cases - if whoever is writing the article doesn’t account for the fact that English uses two concepts (pig vs. pork) for what Spanish uses one (cerdo = puerco etc.), and assumes the default (“pig”), you’ll end with stuff like *“pig consumption has increased” (i.e. “pork consumption has decreased”). And the abstraction layer has no way to know if the human is talking about some living animal or its flesh.
And context doesn’t help much because pork and pigs are mentioned often in the same articles.
If it did matter, you’d explicitly describe the concept of “a living pig with its flesh” instead of the more generic concept of a living pig.
As I said in the top, you’ll end with a “map” that is as large as the “terrain”, thus useless. (Or: spending way more effort explicitly describing all concepts that it’s simply easier to translate it by hand.)
The project isn’t useless, mind you. Perhaps not surprisingly, it could be usable for small things in highly controlled situations, like tables; OP themself hinted this usage.
But as much as I avoid doing “hard” statements about future tech, I’m fairly certain that it won’t be viable as a way to write full articles in a language-agnostic way.
The writer will need to tag things down, to minimal details, for the sake of languages that they don’t care about.
Sure and that’s likely a good bit of work.
However, you must consider the alternative which is translating the entire text to dozens of languages and doing the same for any update done to said text. I’d assume that to be even more work by at least one order of magnitude.
Many languages are quite similar to another. An article written in the hypothetical abstract language and tuned on an abstract level to produce good results in German would likely produce good results in Dutch too and likely wouldn’t need much tweaking for good results in e.g. English. This has the potential to save ton of work.
This issue affects languages as a whole, and sometimes in ways that you can’t arbitrate through a fixed writing style because they convey meaning.
The point of the abstract language would be to convey the meaning without requiring a language-specific writing style. The language-specific writing style to convey the specified meaning would be up to the language-specific “renderers”.
(For example: if you don’t encode the social gender into the 3rd person pronouns, English breaks.)
That’s up to the English “renderer” to do. If it decides to use a pronoun for e.g. a subject that identifies as male, it’d use “he”. All the abstract language’s “sentence” would contain is the concept of a male-identifying subject. (It probably shouldn’t even encode the fact that a pronoun is used as usage of pronouns instead of nouns is also language-specific. Though I guess it could be an optional tag.)
Often there’s no such thing as the “default”. The example with pig/pork is one of those cases - if whoever is writing the article doesn’t account for the fact that English uses two concepts (pig vs. pork) for what Spanish uses one (cerdo = puerco etc.), and assumes the default (“pig”), you’ll end with stuff like *“pig consumption has increased” (i.e. “pork consumption has decreased”). And the abstraction layer has no way to know if the human is talking about some living animal or its flesh.
No, that’d simply be a mistake in building the abstract sentence. The concept of a pig was used rather than the concept of edible meat made from pig which would have been the correct subject to use in this sentence.
Mistakes like this will happen and I’d even consider them likely to happen but the cool thing here is that “pig consumption has increased”, while obviously slightly wrong, would still be quite comprehensible. That’s an insane advantage considering that this would apply to any language for which a generic “renderer” was implemented.
It ends like that story about a map so large that it represents the terrain accurately being as big as the terrain, thus useless.
As I said in the top, you’ll end with a “map” that is as large as the “terrain”, thus useless. (Or: spending way more effort explicitly describing all concepts that it’s simply easier to translate it by hand.)
I don’t see how that would necessarily be the case. Most sentences on Wikipedia are of descriptive nature and follow rather simple structures; only complicated further for the purpose of aiding text flow. Let’s take the first sentence of the Wikipedia article on Lemmy:
Lemmy is a free and open-source software for running self-hosted social news aggregation and discussion forums.
This could be represented in a hypothetical abstract sentence like this:
(explanation (proper-noun "lemmy") (software-facilitating :kind FOSS :purpose (purposes (apply-property 'self-hosted '(news-aggregation-platform discussion-forum)))))
(IDK why I chose lisp to represent this but it felt surprisingly natural.)
What this says is that this sentence explains the concept of lemmy by equating it with the concept of a software which facilitates the combination of multiple purposes.
A language-specific “renderer” such as the English one would then take this abstract representation and turn it into an English sentence:
The concept of an explanation of a thing would then be turned into an explanation sentence. Explanation sentences depend on what it is that is being explained. In this case, the subject is specifically marked as a proper noun which is usually explained using a structure like “<explained thing> is <explanation>”. (An explanation for a different type of word could use a different structure.) Because it’s a proper noun and at the beginning of a sentence, “Lemmy” would be capitalised.
Next the explanation part which is declared as a concept of being software of the kind FOSS facilitating some purpose. The combined concept of an object and its purpose is represented as “<object> for the purpose of <purpose>” in English. The object is FOSS here and specifically a software facilitating some purpose, so the English “renderer” can expand this into “free and open-source software for the purpose of facilitating <purpose>”.
The purpose given is the purpose of having multiple purposes and this concept simply combines multiple purposes into one.
The purposes are two objects to which a property has been applied. In English, the concept of applying a property is represented as as “a <property as adjective> <object>”, so in this case “a self-hosted news-aggregation platform” and “a self-hosted online discussion forum”. These purposes are then combined using the standard English method of combining multiple objects which is listing them: “a self-hosted news-aggregation platform and a self-hosted online discussion forum”. Because both purposes have the same adjective applied, the English “renderer” would likely make the stylistic choice of implicitly applying it to both which is permitted in English: “a self-hosted news-aggregation platform and online discussion forum”.It would then be able to piece together this English sentence: “Lemmy is a free and open source software for the purposes of facilitating a self-hosted news-aggregation platform and online discussion forum.”.
You could be even more specific in the abstract sentence in order to get exactly the original sentence but this is also a perfectly valid sentence for explaining Lemmy in English. All just from declaring concepts in an abstract way and transforming that abstract representation into natural language text using static rules.
The writer will need to tag things down, to minimal details, for the sake of languages that they don’t care about.
Sure and that’s likely a good bit of work.
It isn’t just “a good bit of work”, it’s an unreasonably large amount of work. It’s like draining the ocean with a bucket. I’m talking about tagging hundreds of subtle distinctions for each sentence, and that not tagging those distinctions will output nonsense for at least some language.
However, you must consider [implied: “you didn’t consider”] the alternative which is translating the entire text to dozens of languages
I did consider it. And it’s blatantly clearly overall less work, and easier to distribute among multiple translators.
For example. If I’m translating some genitive construction from Portuguese to Latin, I don’t need to care on which side of English’s esoteric “of vs. 's” distinction it lies in. Or if I’m expected to use の/no in Japanese in that situation. Or to tag “hey, this is not alienable!” for the sake of Nahuatl. I need to deal with oddities of exactly two languages - source and target.
Under the proposed system though? Enjoy tagging a single word [jap-no][eng-of][lat-gen][nah-inal]. And that’s only for four languages.
(inb4: this shit depends on meaning, so no, code can’t handle it. At most code can convert sea[lat-gen] to “maris”, but it won’t “magically” know if it needs to use the genitive or ablative, or if English would use “of” or “'s”.)
and doing the same for any update done to said text
False dichotomy.
I’d assume
If you’re eager to assume (i.e. to make shit up and take it as true), please do not waste my time.
that to be even more work by at least one order of magnitude.
Source: you made it up.
Many languages are quite similar to another. An article written in the hypothetical abstract language and tuned on an abstract level to produce good results in German would likely produce good results in Dutch too and likely wouldn’t need much tweaking for good results in e.g. English. This has the potential to save ton of work.
Okay… I’ve stopped reading here. If your low-hanging fruit example is three closely related languages, then it’s blatantly clear that you’re ignorant on the sheer scale of the problem.
This is an encyclopedia, so there are no pronouns like “I”, so this simplifies this issue. The remaining ones are in the third person, and if we link them to data about the person that is referred to it would solve this. A longuist doesn’t necessarily need to know a language in order to analyze its grammar, and a lot of the work needed in Wikifunctions is like this.
This is an encyclopedia, so there are no pronouns like “I”, so this simplifies this issue. The remaining ones are in the third person, and if we link them to data about the person that is referred to it would solve this.
The pronoun is an example. You are confusing the example with the issue.
This issue is that, if some language out there marks a distinction, whoever writes the abstract version of the text will need to mark it, as that info won’t “magically” pop out of nowhere. The issue won’t appear just in the pronouns, but every where.
A longuist [linguist] doesn’t necessarily need to know a language in order to analyze its grammar, and a lot of the work needed in Wikifunctions is like this.
Usually when you aren’t proficient in a language but still describing it, you focus on a single aspect of its grammar (for example, “unergative verbs”) and either a single variety or a group of related ones.
What the abstract version of the text would require is nowhere close to that. It’s more like demanding the linguist to output a full grammar, to usable levels, of every language found in Wikipedia, to write down a text about some asteroid, using a notation that is cross-linguistically consistent and comprehensible.
Also note that descriptions coming from linguists who are not proficient in a variety in question tend to be poorer.
They’re just going to write all the articles in lojban.
Not even that would do the trick - practical usage of Lojban heavily relies on fu’ivla, that carry with themselves the semantic scope assigned to the original words. .u’i I want to see them trying though.
I’ll reply to myself to highlight a point, and issue a challenge for those who assume that WMF’s apparent goal - to translate Wikipedia articles to more languages by writing them in code that has a lot of linguistic information - is actually viable:
Here’s an excerpt from an actual Wikipedia article: “the solubility of these gases depending on the temperature and salinity of the water.” Show me all the linguistic information that a writer would need to input, to convey the same information, in that system idealised by the goal, in a way that would not output “then who was phone?” tier nonsense for some languages. Then I’ll show you why it would still output nonsense for some languages.
Too much work? Then feel free to do it just for “of the water”. It’s a single PP, how hard would it be? /s
Hic Rhodes, hic salta.
[Edit reason: clarity.]
Your description doesn’t seem to match what the site does? For example the front page has a function that converts uppercase text to lowercase text.
It’s not article content - it’s an interactive utility.
This sounds like more roboticication of wikipedia. Not good.