Distributed meanings and senses in error-driven learning framework – a proof of concept
Conference object (Published version)
Metadata
Show full item recordAbstract
driven learning framework (Filipović Đurđević & Kostić, 2021; Mišić & Filipović Đurđević,
2021). Filpović Đurđević and Kostić (2021) introduced a hypothesis that lexical ambiguity could
be operationalized via partial overlap of multiple cues/outcomes related to meaning. Their
demonstration relied on distributional semantics, namely the co-occurrence of words in the
context. However, although relying on natural language samples is a powerful approach, it also
introduces many complexities that potentially obscure the learning mechanics behind the
ambiguity effects. Therefore, our aim was to perform ambiguity learning simulations from more
of a theoretical standpoint by employing the toy model approach.
Theory informed our data generation process in two ways. First, error-driven learning
(Rescorla & Wagner, 1972) offered a mechanism for learning ambiguous words and the
importance of cue competition for learning to occur (Hoppe et al., 2022). Second, we relied on
psycholinguisti...c theory for descriptions of ambiguity, namely the polysemy (multiple related
senses) and homonymy (multiple unrelated meanings) distinction (Rodd et al., 2002). In addition
to sense/meaning relatedness, we also paid attention to their probabilities (Filipović Đurđević &
Kostić, 2017). We manipulated the type of lexical ambiguity (unambiguous words, polysemes,
homonyms), the balance of sense/meaning probabilities (balanced, unbalanced), and the level
of cue competition (low, medium, high).
Data were generated in the following way. We modelled a total of six words: two
unambiguous words, a balanced and an unbalanced polyseme, and a balanced and an
unbalanced homonym. Each word was represented by one outcome. Cues were created
separately for each sense/meaning and were constructed as an equal-length string of arbitrary
elements. The ambiguity type was manipulated via the cue overlap. Unambiguous words were
predicted by a single cue set. Homonyms were predicted by three distinct sets of cues, each
representing one meaning. Polysemous words were also predicted by three sets of cues,
however, in addition to some unique cues for each of the senses, sets had some overlap among
themselves in order to represent the sense relatedness. Each of the artificial words (outcomes)
and its cues was presented to the network an equal amount of times. Balance of the
sense/meanings frequency distribution was manipulated through the frequency of the
presentation of each cue-outcome pairing. Finally, to introduce more cue competition, we
randomly sampled a number of existing cues and appended them to other meanings/senses
strings. By varying the number of cues appended, we varied the cue competition intensity. The
data structure scheme is presented in Figure 1.
We then compared simulations on two different measures – the activation of the
outcomes, and the learnability (a quantitative description of learning curves). When cue
competition was present, activation decreased in the following order: balanced homonyms,
unbalanced homonyms, unbalanced polysemes, and unbalanced polysemes, with unambiguous
words, activated the least. This pattern, although expected to be inversely proportionate, was
directly proportionate to the RTs in lexical decision tasks (Filipović Đurđević, 2019; Filipović
Đurđević & Kostić, 2021). Learnability measure revealed that homonyms were learned the best,
followed by polysemes, and then unambiguous words. Nevertheless, the existing relationship
suggests that possible modifications of the generated data might lead to a better insight into
how learning leads to the presence of ambiguity in language.
Keywords:
distributed meanings / distributed senses / error-driven learningSource:
Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany, 2022, 12-13Funding / projects:
- Ministry of Science, Technological Development and Innovation of the Republic of Serbia, institutional funding - 200163 (University of Belgrade, Faculty of Philosophy) (RS-MESTD-inst-2020-200163)
Institution/Community
Psihologija / PsychologyTY - CONF AU - Mišić, Ksenija AU - Filipović Đurđević, Dušica PY - 2022 UR - http://reff.f.bg.ac.rs/handle/123456789/5135 AB - driven learning framework (Filipović Đurđević & Kostić, 2021; Mišić & Filipović Đurđević, 2021). Filpović Đurđević and Kostić (2021) introduced a hypothesis that lexical ambiguity could be operationalized via partial overlap of multiple cues/outcomes related to meaning. Their demonstration relied on distributional semantics, namely the co-occurrence of words in the context. However, although relying on natural language samples is a powerful approach, it also introduces many complexities that potentially obscure the learning mechanics behind the ambiguity effects. Therefore, our aim was to perform ambiguity learning simulations from more of a theoretical standpoint by employing the toy model approach. Theory informed our data generation process in two ways. First, error-driven learning (Rescorla & Wagner, 1972) offered a mechanism for learning ambiguous words and the importance of cue competition for learning to occur (Hoppe et al., 2022). Second, we relied on psycholinguistic theory for descriptions of ambiguity, namely the polysemy (multiple related senses) and homonymy (multiple unrelated meanings) distinction (Rodd et al., 2002). In addition to sense/meaning relatedness, we also paid attention to their probabilities (Filipović Đurđević & Kostić, 2017). We manipulated the type of lexical ambiguity (unambiguous words, polysemes, homonyms), the balance of sense/meaning probabilities (balanced, unbalanced), and the level of cue competition (low, medium, high). Data were generated in the following way. We modelled a total of six words: two unambiguous words, a balanced and an unbalanced polyseme, and a balanced and an unbalanced homonym. Each word was represented by one outcome. Cues were created separately for each sense/meaning and were constructed as an equal-length string of arbitrary elements. The ambiguity type was manipulated via the cue overlap. Unambiguous words were predicted by a single cue set. Homonyms were predicted by three distinct sets of cues, each representing one meaning. Polysemous words were also predicted by three sets of cues, however, in addition to some unique cues for each of the senses, sets had some overlap among themselves in order to represent the sense relatedness. Each of the artificial words (outcomes) and its cues was presented to the network an equal amount of times. Balance of the sense/meanings frequency distribution was manipulated through the frequency of the presentation of each cue-outcome pairing. Finally, to introduce more cue competition, we randomly sampled a number of existing cues and appended them to other meanings/senses strings. By varying the number of cues appended, we varied the cue competition intensity. The data structure scheme is presented in Figure 1. We then compared simulations on two different measures – the activation of the outcomes, and the learnability (a quantitative description of learning curves). When cue competition was present, activation decreased in the following order: balanced homonyms, unbalanced homonyms, unbalanced polysemes, and unbalanced polysemes, with unambiguous words, activated the least. This pattern, although expected to be inversely proportionate, was directly proportionate to the RTs in lexical decision tasks (Filipović Đurđević, 2019; Filipović Đurđević & Kostić, 2021). Learnability measure revealed that homonyms were learned the best, followed by polysemes, and then unambiguous words. Nevertheless, the existing relationship suggests that possible modifications of the generated data might lead to a better insight into how learning leads to the presence of ambiguity in language. C3 - Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany T1 - Distributed meanings and senses in error-driven learning framework – a proof of concept EP - 13 SP - 12 UR - https://hdl.handle.net/21.15107/rcub_reff_5135 ER -
@conference{ author = "Mišić, Ksenija and Filipović Đurđević, Dušica", year = "2022", abstract = "driven learning framework (Filipović Đurđević & Kostić, 2021; Mišić & Filipović Đurđević, 2021). Filpović Đurđević and Kostić (2021) introduced a hypothesis that lexical ambiguity could be operationalized via partial overlap of multiple cues/outcomes related to meaning. Their demonstration relied on distributional semantics, namely the co-occurrence of words in the context. However, although relying on natural language samples is a powerful approach, it also introduces many complexities that potentially obscure the learning mechanics behind the ambiguity effects. Therefore, our aim was to perform ambiguity learning simulations from more of a theoretical standpoint by employing the toy model approach. Theory informed our data generation process in two ways. First, error-driven learning (Rescorla & Wagner, 1972) offered a mechanism for learning ambiguous words and the importance of cue competition for learning to occur (Hoppe et al., 2022). Second, we relied on psycholinguistic theory for descriptions of ambiguity, namely the polysemy (multiple related senses) and homonymy (multiple unrelated meanings) distinction (Rodd et al., 2002). In addition to sense/meaning relatedness, we also paid attention to their probabilities (Filipović Đurđević & Kostić, 2017). We manipulated the type of lexical ambiguity (unambiguous words, polysemes, homonyms), the balance of sense/meaning probabilities (balanced, unbalanced), and the level of cue competition (low, medium, high). Data were generated in the following way. We modelled a total of six words: two unambiguous words, a balanced and an unbalanced polyseme, and a balanced and an unbalanced homonym. Each word was represented by one outcome. Cues were created separately for each sense/meaning and were constructed as an equal-length string of arbitrary elements. The ambiguity type was manipulated via the cue overlap. Unambiguous words were predicted by a single cue set. Homonyms were predicted by three distinct sets of cues, each representing one meaning. Polysemous words were also predicted by three sets of cues, however, in addition to some unique cues for each of the senses, sets had some overlap among themselves in order to represent the sense relatedness. Each of the artificial words (outcomes) and its cues was presented to the network an equal amount of times. Balance of the sense/meanings frequency distribution was manipulated through the frequency of the presentation of each cue-outcome pairing. Finally, to introduce more cue competition, we randomly sampled a number of existing cues and appended them to other meanings/senses strings. By varying the number of cues appended, we varied the cue competition intensity. The data structure scheme is presented in Figure 1. We then compared simulations on two different measures – the activation of the outcomes, and the learnability (a quantitative description of learning curves). When cue competition was present, activation decreased in the following order: balanced homonyms, unbalanced homonyms, unbalanced polysemes, and unbalanced polysemes, with unambiguous words, activated the least. This pattern, although expected to be inversely proportionate, was directly proportionate to the RTs in lexical decision tasks (Filipović Đurđević, 2019; Filipović Đurđević & Kostić, 2021). Learnability measure revealed that homonyms were learned the best, followed by polysemes, and then unambiguous words. Nevertheless, the existing relationship suggests that possible modifications of the generated data might lead to a better insight into how learning leads to the presence of ambiguity in language.", journal = "Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany", title = "Distributed meanings and senses in error-driven learning framework – a proof of concept", pages = "13-12", url = "https://hdl.handle.net/21.15107/rcub_reff_5135" }
Mišić, K.,& Filipović Đurđević, D.. (2022). Distributed meanings and senses in error-driven learning framework – a proof of concept. in Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany, 12-13. https://hdl.handle.net/21.15107/rcub_reff_5135
Mišić K, Filipović Đurđević D. Distributed meanings and senses in error-driven learning framework – a proof of concept. in Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany. 2022;:12-13. https://hdl.handle.net/21.15107/rcub_reff_5135 .
Mišić, Ksenija, Filipović Đurđević, Dušica, "Distributed meanings and senses in error-driven learning framework – a proof of concept" in Proceedings of the Second International Conference on Error-Driven Learning in Language (EDLL 2022), August 1-3, University of Tübingen, Germany (2022):12-13, https://hdl.handle.net/21.15107/rcub_reff_5135 .