Next Article in Journal
On Isotropy of Multimodal Embeddings
Previous Article in Journal
Monkeypox, Disinformation, and Fact-Checking: A Review of Ten Iberoamerican Countries in the Context of Public Health Emergency
Previous Article in Special Issue
Multi-Task Romanian Email Classification in a Business Context
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Findings on Ad Hoc Contractions

Department of Computer Science, University of Nevada, Las Vegas, NV 89154, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Information 2023, 14(7), 391; https://doi.org/10.3390/info14070391
Submission received: 18 May 2023 / Revised: 24 June 2023 / Accepted: 25 June 2023 / Published: 10 July 2023
(This article belongs to the Special Issue Novel Methods and Applications in Natural Language Processing)

Abstract

:
Abbreviations are often overlooked, since their frequency and acceptance are almost second nature in everyday communication. Business names, handwritten notes, online messaging, professional domains, and different languages all have their own set of abbreviations. The abundance and frequent introduction of new abbreviations cause multiple areas of overlaps and ambiguity, which mean documents often lose their clarity. We reverse engineered the process of creating these ad hoc abbreviations and revealed some preliminary statistics on what makes them easier or harder to define. In addition, we generated candidate definitions for which it proved difficult for a word sense disambiguation model to select the correct definition.

1. Introduction

Abbreviations have widespread, everyday usage across multiple languages and domains. Their ease of use and time-saving benefits have incentivized their utilization in both professional and casual contexts. An analysis by Barnett and Doubleday was performed on the abstracts and titles of scientific literature from the years 1950 to 2019 on the usage of acronyms. Their results showed an increase in the use of acronyms from 0.7 per 100 words in 1950, to 2.4 per 100 words in 2019 [1].
The increasing amount of abbreviations in the professional sciences alone causes clarity issues and misunderstandings when trying to interpret the long form definitions. A survey done by Sheppard et al. [2] found over 2286 abbreviations being used in 25 clinical handout sheets. According to Tariq and Sharma, within the last twenty years, approximately 7000 to 10,000 people have died due to medical mistakes every year in the United States alone.Among these errors, misunderstanding about abbreviations have contributed to roughly 5% of these casualties. [3]. Minimizing the clarity issue with abbreviations could save a large portion of those documented 5% of lives, and researchers Liu et al. showed that expanding the definitions of abbreviations overwhelmingly increased the patient comprehension in regards to their health records [4].
While abbreviations can pose an immediate and fatal danger in the medical field, ambiguity regarding them can plague every text-related field, including computer science. Source code utilizes many creative shortenings of words and phrases to facilitate typing speed. Deployment of incorrect abbreviations or their over-usage can severely hurt the readability and understanding [5] of code. While researchers Jiang et al. found that overly lengthy words in code were detrimental to reader comprehension, and recommended the frequent usage of abbreviations to improve understanding, Hales et al. made the argument that this leads to more instances of misuse. This, when paired with the usage of esoteric jargon, can lead to readers becoming alienated and overwhelmed with the text, increasing confusion and misunderstandings [6].
First, we must lay a foundation of what acronyms and abbreviations are, before modifying their properties to satisfy their professional usage. Acronyms are usually a type of abbreviation where a phrase or sequence of words are shortened into the initials of each major word (e.g., DOE stands for Department of Energy). An abbreviation is generally a relative term used to describe the shortening of a word using contractions or contracting parts of the word (e.g., dept stands for department). Each type of abbreviation will often follow a loose set of rules and cues that hint the reader that it is an abbreviation. Taghva and Gilbreth created the first rule-based machine algorithm to detect acronyms. This algorithm operated under rules of identification where the acronym length must be at least three letters long and its full nomenclature or definition must be provided within a certain window of text [7]. This later developed into a machine learning approach, utilizing HMMs (Hidden Markov Models), to accurately select the definition within a proportionate window size [8]. In a later section, we will modify some of these general purpose rules to better suit the medical domain and to suit contraction-type abbreviations.
Currently, there are three main approaches to handling abbreviations with regards to machine reading comprehension. The first approach involves rule-based/statistical models, such as HMMs or decision trees, which require both understanding the cues and inferences of the document, to select the best definition. These approaches often do well when data are structured and uniform, but fail in real-life scenarios, because abbreviations do not have a uniform set of rules that every author follows. In particular, these methods require some sort of regulation of their usage, such as employing all capital letters or including a definition in parentheses to indicate the existence of an acronym. Certain models use prior knowledge of context to determine the best expansion or try to find its definition nearby.
The second approach is through maximizing word sense disambiguation (WSD). This approach typically utilizes the context to determine the meaning from a set of potential definitions and selects the one which makes the most sense. Researchers such as Sultan et al. have considered parts-of-speech (POS) to determine whether or not definitions fit their grammatical place. In addition to their usage of POS, the authors also created pipelines of alignment modules to determine how closely related certain words are, based on their syntactic dependencies [9]. Other researchers working on WSD thought to use graph-like approaches such as PageRank [10,11] to map relationships between pages and words. This later developed further, as researchers started to correlate word meanings as vertices and the edges as the semantic relationships derived from WordNet to construct a disambiguation graph [12]. The disambiguation graph provides a context and understanding of words in relation to each other and can offer a deeper insight into the likelihood of determining which words fit which contexts. Similarly to PageRank, a different approach of considering the term frequency-inverse document frequency (TF-IDF) was utilized to relate context with frequency of occurrence [13,14]. TF-IDF is a common indexing method used for analyzing likelihood and has applications in areas such as web page retrieval and recommendation systems but can also be utilized to analyze text and the context of words. The authors Turtel and Shasha utilized TF-IDF for acronym disambiguation [15]. Similarly, the authors Li et al. utilized word embedding paired with TF-IDF to solve acronym disambiguation [16].
The third approach is to use a transformer model such as BERT (bidirectional encoder representations from transformers) [17,18]. Traditionally, recurrent neural networks using a long short-term memory have been deployed for natural language models. Recently BERT has replaced most of those networks. Essentially, BERT is an upgrade on recurrent neural network models, as it is able to take in information directionlessly. Previously, we demonstrated some of the power of BERT in its ability to generate abbreviation definitions [19] and further successfully tested it on ambiguous definitions [13]. The authors Daza et al. utilized a SloBERTa model with an additional single neural layer to tackle abbreviation disambiguation for Slovenian biographical lexicons [20]. The benefits of using BERT models include having access to differently trained variations on domain-specific data including, but not limited to, models such as RoBERTa [21], SciBERT [22], and AlBERT [23].
Within this scope, there are still many limitations that can be improved upon. The most salient problem is the lack of domain specific data and centralized data. In regards to proprietary and personal data, especially in the medical field, there is a lack of sufficient datasets that encompass all definitions and their multiple abbreviation forms. Additionally, ambiguous definitions can exist across domains, with no accounting for these disparities. There has been some progress towards creating a consolidation of publicly available data, such as the database MeDal, which is composed of 14,393,619 medical articles and abstracts. The authors, Wen et al. utilized this dataset to pre-train several machine learning models. The results showed improvements towards being able to accurately define abbreviations [24]. Alternatively, the researchers Skreta et al. combined data sampling and reverse sampling (RS) to automatically create their dataset, without the need for human aid [25].
Abbreviations, with their convenience of use and their improvement of efficiency in all text-related fields, will continue to find prolific usage in professional contexts. Inevitably, this leads to their misunderstanding and misinterpretation, due to definition ambiguity and over-utilization. However, with insight into how abbreviations are commonly used and keeping to a standard of abbreviation guidelines, we could minimize these potential misreadings. In this paper, we introduce a novel way to generate ad hoc abbreviations, produce their definitions, and to reverse engineer their candidate definitions. Our work is still in the early stages, but we have found several interesting statistics, such as higher definition retrieval rates for abbreviations retaining 40% of their characters and trends such as a 93% omission rate for vowels. These preliminary results can be further utilized to lay foundations and create certain rules involving contraction-type abbreviations.

2. Materials and Methods

To fully tackle abbreviation disambiguation, the ideal scenario is to have a dictionary with all possible abbreviations and their definitions assigned. Realistically, this is not plausible, as new abbreviations and deviations are constantly being produced. However, it may be possible to generate all or close to all possible abbreviations that are likely to be used and assign them sufficient context to aid WSD and allow selecting the best possible candidate for expansion.
Currently, most of the work and data have focused on tackling acronyms. However, not much work has been done around contraction-based abbreviations. Contractions are a subset of abbreviations where, instead of a phrase being abbreviated, a single word is. Similarly to acronyms, contraction-type abbreviations can be related to sub-string matching, but with characters. Pennel et al. introduced a normalization model for contractions found in Twitter data [26]. Their method leans more towards a statistical model for detection and focuses on manual annotations to create reasonable abbreviations.
Traditionally, acronyms have unique characteristics, such as containing all or a majority of capital letters, annotated within parentheses, and having long-form definitions provided. Contractions tend to have traits that are more difficult for a computer to parse, such as having punctuation as part of the word, containing misspellings of words, and having a collection of letters that do not make for fluid pronunciation. Additionally, what makes this type of abbreviation harder to expand is the lack of a provided definition. The underlying goal of contraction expansion is to identify which long-form definition the substring could soundly be a part of.
Figure 1 shows our experimental design process. In the later sections, we will elaborate more on the data generation process, word selection, abbreviation process, and reconstruction of the definitions.

2.1. Dataset

In order to simulate a larger amount of domains and introduce ambiguity, we wanted to create a corpus that introduces different topics. Initially, we intended to use the MeDal dataset, but the provided dataset contained a vast amount of errors, including typos and unidentifiable characters and symbols. Instead, we collected different text files, generated from Wikipedia articles. These were text documents of random Wikipedia pages, created by using the built in random article generator in the Wikipedia API for python. We simulated different domains by manually curating articles to have a diversity of domains and genres, with some small overlaps. Our final dictionary contained 35 different articles in various domains, including but not limited to autobiographies, education, music, technology (circuits), and film. There was no strict criteria for document acceptance, other than a soft attempt to maximize domain and genre diversity. We would like to argue that contraction-based abbreviations can be used with any part of speech (noun, verb, etc.), which is why we did not use a stricter method for accepting documents. In addition, document size did not play a role in our selection process, as we wanted to keep the simulated variance.
We then calculated the number of occurrences of each word and the document domain they occurred in. We kept everything in the corpus as it was presented, and introduced a few rules to select abbreviation candidates to be contracted. Let us denote the candidate abbreviation set as C.
For candidate words, we make a few modifications to our criteria:
  • Frequency—the amount of times the word that is intended to be abbreviated is used; it should be quite frequent to warrant the need to abbreviate. We decided on a minimum of seven occurrences;
  • Length—the length of the candidate word must be greater than five letters;
  • Outliers—unnatural words such as words with numbers were not included;
  • Replacers—abbreviations with replacement characters or symbols that were not found in the original long-form definition (ex: XFER with X in place of the word Trans) were omitted.
To further elaborate our justifications for each rule, we chose words with frequencies greater than seven occurrences, as this meant the term could potentially be used in more than one domain and yielded a higher contextual significance. On average, our occurrence rate was above 10 uses for words that could be found in at least 2 documents. We deliberately did not choose any stop words, as they are generally uninformative [27] and have commonly understood or established contractions already. We also excluded words that had a length less than five letters, as they would not require abbreviation. Our average word length was approximately eight letters. Often, in medical domains or fields with high Latin and Greek usage, such as chemistry, compounds and scientific names are abbreviated using numbers and Greek letters. We chose to omit these types of words as they have more domain specific rules and guidelines that cannot be generalized easily. Lastly, for candidate definitions, we chose to omit any symbolic representations for similar reasons, as they have their own specific guidelines to follow.

2.2. Ad Hoc Abbreviations

Since abbreviations have no strict rules for generation, different variations of the same word can occur. We wanted to investigate what characteristics may impact the generation of ad hoc abbreviations and whether or not guidelines can be established to promote better abbreviation generation. For each item in C, we utilized an online user submitted dictionary, https://www.abbreviations.com/, (accessed: 24 April 2023) to find the most common abbreviations for each word. We chose the top-five unique abbreviations for each item and created a dictionary D for them. It is important to note that not every definition had more than five unique abbreviations to choose from. Table 1 shows a subset of a long-form word and its candidate ad hoc abbreviations.
From our various contractions, we needed to assign candidate definitions to them. To simulate potential ambiguous definitions, we needed to perform a dictionary look up for all the words containing the contraction letters. We utilized an online platform https://www.dcode.fr/en (accessed: 24 April 2023). This allowed us to input any substring and search a dictionary to find all words containing the substring.
From the returned substrings, we removed strings that did not start with the same letter as the abbreviation and randomly selected five potential definitions from the remainder. Among the five selected definitions, one was always the ground truth definition. If a substring that was selected had the same root, such as a plural form or different tense, we removed it and randomly chose a different word, if available. Additionally, not all abbreviations contained more than five candidate definitions. We noticed a positive correlation between the increase in substring length and the decrease in potential candidates. Table 2 shows a sample ad hoc abbreviation found in Table 1 and its potential long-form expansions using a dictionary look up.
After building up a corpus of potential abbreviations and their candidate long forms, we tested for ambiguity utilizing a TF-IDF model and BERT. However, in practice, the BERT model yielded poor results, largely due to the lack of definitions retained in the model. In a previous paper [19], we explored the use of BERT on acronym-type abbreviations, and it presented a 94% acceptance rate. Using the same method, but querying for contractions, the BERT model scored 0% for acceptance.
There are two major flaws with using the BERT model. The first issue is that the document containing the ground truth must be given to the model. Outside of training, BERT will not retain new information. The second issue is the model itself does not comprehend contractions. For example, when prompted for acronyms such as “What does RAM mean?” the model will return random access memory. However, when prompted for contractions, an example prompt would include “What does bg mean?” and its corresponding result would often return junk or an entire snippet of the document such as “TECHNICAL FIELD The present invention relates to a wireless local access network (WLAN), and more particularly, to a channel access mechanism for a very high throughput (VHT) WLAN system and a station supporting the channel access mechanism. BACKGROUND ART …”.
For the remainder of our WSD for definition expansion, we will use TF-IDF. Since each document can be roughly treated as its own category or domain (after curation), the WSD assumption would correlate with the correct definition in its respective domain.

3. Results

Our dictionary lookup utilizing https://www.dcode.fr/en generated over 1000 candidates. Table 3 details the different candidate definitions chosen, as well as the number of potential candidate definitions to chose from, what the ratio of removed letters were in the contraction, and statistics on the vowel removal. We noticed a drop in the number of potential candidates to choose from, depending on the ratio of the length of the substring to the actual word. Naturally, the more characters that are retained, the less likely it can generate extra options that can be ambiguous. To summarize, we found that, on average, utilizing 50.769% of the characters to abbreviate reduced the candidate options from over 1000, to less than 100. The lowest occurrence we found was when utilizing 40% of a string reduced candidates to 53.
Out of the 58 vowels that were found in the long-form definitions, not including the starting letter, we found that 93.103% of vowels were removed by the ad hoc abbreviations. An interesting statistic to mention is that none of the abbreviations’ low candidate selections kept any non-first-letter vowels. This may imply that vowels contain a low amount of information in contraction-type abbreviations. Conceptually, this makes sense, as there are usually only five potential vowels to chose from.
Testing for ambiguity between definitions, we calculated the TF-IDF for each definition per category. Candidate definitions that do not appear in our corpus scored 0. We have provided the link to our TF-IDF calculations, as well as the curated text files used to build our dictionary (https://github.com/choivsh/contraction-data). Among our abbreviations, we found six abbreviations that had ambiguous definitions. These six abbreviations included adj, ad, mods, pry, py, and prim. Examples of their ambiguous definitions included words such as modules, modification, primary, and proxy.
We were only interested in whether or not TF-IDF was strong enough to disambiguate the definitions, so we compared the document probabilities of the overlapped data. Essentially, any document that did not overlap would guarantee the ground truth definition was the only contender. For the overlapping documents, we found that only one out of seven categories selected the correct definition, but this was significant (the difference in probability was greater than 0.01). For the remaining six categories that were marked incorrect, the probability difference was less than 0.01, making the results lean towards being inconclusive. Table 4 is a subset of the probability calculations based on our TF-IDF model.

4. Discussion

Overall, this study can be generalized as a proof of concept of creating a dictionary containing various ad hoc abbreviations. The results in our sample demonstrated some causal correlations between which letters can be truncated while retaining enough information to narrow the definition search further. If coupled with categorical tagging, machines are able to disseminate and replace contraction-type abbreviations with their long-form definitions. We would like to discuss some improvements, as well as future work that this line of research could be extended to.
First, we would like to explore some limitations of our model. The candidate definitions that came from the dictionary lookup yielded over 1000 visible entries and permutations. As there was no available API for https://www.dcode.fr/en, we did not have an effective way of reducing the candidate definitions by removing any definition that did not start with the same starting character (e.g., bg could return: airbag). This may artificially inflate the number of candidates returned; however, the overall trend of increasing substring characters would still reduce the amount of candidates returned. A potential solution would be to store a standardized English dictionary and query with the criteria needed to decipher.
Second, we did not have an effective way of automatically generating ad hoc contractions. While it is possible to permute different iterations of the candidate string, we argue that having real-life data and utilizing user submitted abbreviations more accurately simulates an ad hoc nature. Utilizing this method, we could lay a foundation and extract a few important rules and assumptions based on the findings, to implement a model in the future that reduces the amount of nuisance and noisy candidate abbreviations. However, an argument could be made to generate those permutations, to see whether or not a pattern could be discovered that performs better than the naturally accepted abbreviations made by humans.
Following this train of thought, we could build a rule-based model to anticipate and predict the probability of the next letter in the substring and reverse engineer the best candidate contraction for each word. This could be closely related to a HMM used for acronyms. In future work, we plan to use further pattern recognition techniques to extract more characteristics that make up a contraction and determine a stricter set of rules, to reduce the amount of ambiguity in definition finding. Currently, we have established that maintaining 40% of the letters in a contraction-type abbreviation may be a good baseline to filter candidate definitions. We are considering investigating a word’s features, such as the context and part of speech, as this may give indications that can eliminate potential candidates. This would allow for less strict guidelines, in the hope that it is possible to reduce the ratio of letters needed.
One additional note is the relative size of our dataset. We have demonstrated as a proof of concept that we can build a small dictionary of terms, and this model could be scaled further, to create a larger corpus for better results. Context is usually given to abbreviations, making their domain typically known. Specialized dictionaries could be made to help target more specific domain abbreviations. When scaled for generalized abbreviations, a common practice in analyzing TF-IDF values is to take the negative log of the percentages to yield a more easily observed significance.

5. Conclusions

Based on our findings, we were able to artificially simulate some ad hoc abbreviations that could potentially be used in everyday language. From this, we were able to discover some interesting statistics about contraction-type abbreviations, such as vowel usage and removal, and the impact of string length ratio. This area could be further extended to discovering patterns within substrings that can minimize the number of potential candidate definitions. Additionally, we ran tests on these ad hoc abbreviations and generated candidate definitions, which proved to be difficult for the TF-IDF model to evaluate. Our proof of concept shows that contraction-typed abbreviations behave differently than acronyms, and minimizing the potential candidate definitions would be a step in reducing their ambiguity. In future work, we plan to automate the current steps, to scale the concept and uncover more potential patterns for generating easily understood and retrievable contraction-type abbreviations. In addition, we plan to localize this concept and apply it in specific domains, such as the biomedical field, to fully test our findings when the context is much narrower.
Table A1 is a shortened excerpt from the main TF-IDF calculations (the calculations are available on GitHub link provided in later section). Each document category has its own summary of probability (some categories are named, some are numerical, depending on how the API’s metadata named the category). We tallied the total number of occurrences as well. In total we had 35 categories to choose from, with 612 unique words detected in the dictionary.

Author Contributions

Conceptualization, S.C. and K.T.; methodology, S.C. and K.T.; software, S.C.; validation, S.C. and K.T.; formal analysis, S.C.; investigation, S.C. and K.T.; resources, S.C.; data curation, S.C.; writing—original draft preparation, S.C.; writing– review and editing, S.C. and K.T.; visualization, S.C. and K.T.; supervision, K.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Dataset used can be found here at: https://github.com/choivsh/contraction-data.

Acknowledgments

We would like to acknowledge the open resources provided by Wikipedia, https://www.dcode.fr/en, and https://www.abbreviations.com/, which made it possible to generate the data for our experiments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BERTBidirectional Encoder Representations from Transformers
WSDWord Sense Disambiguation
TF-IDFTerm Frequency-Inverse Document Frequency
HMMHidden Markov Models
POSParts of Speech
RSReverse Sampling
APIApplication Program Interface

Appendix A

Table A1. TF-IDF Sample Data.
Table A1. TF-IDF Sample Data.
Document IDDictionary Words
appearancesappearsappendedappletalkappletsappliancesapplicabilityapplicableapplication
10000.00529335200000.0050596380
12000.00908126200000.0043401510
1300.0053297790.0036495730.0059498930.0059498930.011899787000.031551542
14000.002396322000000.007533401
1500000000.0102978470.005080364
16000000000.002266803
17000.00608167900000.0038754390.00637306
18000.005772530000.0021075280.0041382450.002268413
2011_Saransk_Cup000000000
23000000000
40.00194091300.002381053000000.004678373
420000000000
519000000000.096982195
600000000.0237191970.00097514
69900000000.0025445190.018829786
70000000.00280328800.006034572
795000.00125136800000.0023922340.000983492
800000000.0067622650.001010943
9000000000.001361715
Occurrances1281112914

References

  1. Barnett, A.; Doubleday, Z. Meta-Research: The growth of acronyms in the scientific literature. eLife 2020, 9, e60080. [Google Scholar] [CrossRef] [PubMed]
  2. Sheppard, J.E.; Weidner, L.C.E.; Zakai, S.; Fountain-Polley, S.; Williams, J. Ambiguous abbreviations: An audit of abbreviations in paediatric note keeping. Arch. Dis. Child. 2008, 93, 204–206. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Tariq, R.A.; Sharma, S. Inappropriate Medical Abbreviations—Statpearls—NCBI Bookshelf. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2022. Available online: https://www.ncbi.nlm.nih.gov/books/NBK519006 (accessed on 24 April 2023).
  4. Grossman Liu, L.; Russell, D.; Reading Turchioe, M.; Myers, A.C.; Vawdrey, D.K.; Masterson Creber, R.M. Effect of Expansion of Abbreviations and Acronyms on Patient Comprehension of Their Health Records: A Randomized Clinical Trial. JAMA Netw. Open 2022, 5, e2212320. [Google Scholar] [CrossRef] [PubMed]
  5. Jiang, Y.; Liu, H.; Zhang, Y.; Niu, N.; Zhao, Y.; Zhang, L. Which Abbreviations Should Be Expanded? In Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, Athens, Greece, 23–28 August 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 578–589. [Google Scholar] [CrossRef]
  6. Hales, A.H.; Williams, K.D.; Rector, J.H. Alienating the Audience: How Abbreviations Hamper Scientific Communication. APS Obs. 2017, 30. [Google Scholar]
  7. Taghva, K.; Gilbreth, J. Recognizing acronyms and their definitions. Int. J. Doc. Anal. Recognit. 1999, 1, 191–198. [Google Scholar] [CrossRef] [Green Version]
  8. Taghva, K.; Vyas, L. Acronym Expansion Via Hidden Markov Models. In Proceedings of the 2011 21st International Conference on Systems Engineering, Las Vegas, NV, USA, 16–18 August 2011; pp. 120–125. [Google Scholar] [CrossRef]
  9. Sultan, M.A.; Bethard, S.; Sumner, T. Back to Basics for Monolingual Alignment: Exploiting Word Similarity and Contextual Evidence. Trans. Assoc. Comput. Linguist. 2014, 2, 219–230. [Google Scholar] [CrossRef]
  10. Brin, S.; Page, L. The anatomy of a large-scale hypertextual Web search engine. Comput. Netw. ISDN Syst. 1998, 30, 107–117. [Google Scholar] [CrossRef]
  11. Page, L.; Brin, S.; Motwani, R.; Winograd, T. The PageRank Citation Ranking: Bringing Order to the Web; Technical Report 1999-66; Stanford InfoLab: Stanford, CA, USA, 1999. [Google Scholar]
  12. Mihalcea, R.; Tarau, P.; Figa, E. PageRank on Semantic Networks, with Application to Word Sense Disambiguation. In Proceedings of the 20th International Conference on Computational Linguistics, Geneva, Switzerland, 23–27 August 2004; Association for Computational Linguistics: Stroudsburg, PA, USA, 2004; pp. 1126–1132. [Google Scholar] [CrossRef]
  13. Choi, S.; Taghva, K. Abbreviation Disambiguation: A Review of Modern Techniques to Improve Machine Reading Comprehension. In Proceedings of the SAI Computing Conference 2023, London, UK, 22–23 June 2023; Springer Nature: Berlin, Germany, 2023. [Google Scholar]
  14. Song, S.; Miao, Q.; Shi, Z.; Meng, Y.; Chen, H. Co-occurrence semantic knowledge base construction for abbreviation disambiguation. In Proceedings of the 2017 International Conference on Asian Language Processing (IALP), Singapore, 5–7 December 2017; pp. 326–329. [Google Scholar] [CrossRef]
  15. Turtel, B.D.; Shasha, D. Acronym Disambiguation; Technical Report 2007; Courant Institute of Mathematical Sciences Library: New York, NY, USA, 2007. [Google Scholar]
  16. Li, C.; Ji, L.; Yan, J. Acronym Disambiguation Using Word Embedding. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; AAAI Press: Washington, DC, USA, 2015; pp. 4178–4179. [Google Scholar]
  17. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  18. Rush, A.M. The Annotated Transformer. In Proceedings of Workshop for NLP Open Source Software (NLP-OSS); Association for Computational Linguistics: Melbourne, Australia, 2018; pp. 52–60. [Google Scholar]
  19. Choi, S.; Puranik, P.; Dahal, B.; Taghva, K. How to generate data for acronym detection and expansion. Adv. Comput. Intell. 2022, 2, 23. [Google Scholar] [CrossRef]
  20. Daza, A.; Fokkens, A.; Erjavec, T. Dealing with Abbreviations in the Slovenian Biographical Lexicon. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; Association for Computational Linguistics: Abu Dhabi, United Arab Emirates, 2022; pp. 8715–8720. [Google Scholar]
  21. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  22. Beltagy, I.; Cohan, A.; Lo, K. SciBERT: Pretrained Contextualized Embeddings for Scientific Text. arXiv 2019, arXiv:1903.10676. [Google Scholar]
  23. Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; Soricut, R. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. arXiv 2019, arXiv:1909.11942. [Google Scholar]
  24. Wen, Z.; Lu, X.H.; Reddy, S. MeDAL: Medical Abbreviation Disambiguation Dataset for Natural Language Understanding Pretraining. In Proceedings of the 3rd Clinical Natural Language Processing Workshop, Online, 19 November 2020; Association for Computational Linguistics: Abu Dhabi, United Arab Emirates, 2020; pp. 130–135. [Google Scholar]
  25. Skreta, M.; Arbabi, A.; Wang, J.; Brudno, M. Training without training data: Improving the generalizability of automated medical abbreviation disambiguation. arXiv 2019, arXiv:1912.06174. [Google Scholar]
  26. Pennell, D.; Liu, Y. Toward text message normalization: Modeling abbreviation generation. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5364–5367. [Google Scholar] [CrossRef]
  27. Sarica, S.; Luo, J. Stopwords in technical language processing. PLoS ONE 2021, 16, e0254937. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Experiment Flowchart.
Figure 1. Experiment Flowchart.
Information 14 00391 g001
Table 1. Long-form definition with candidate contractions.
Table 1. Long-form definition with candidate contractions.
WordInstancesCandidate 1Candidate 2Candidate 3Candidate 4Candidate 5
background18bgbkgdbkgndbkbgr
additional13addladdtladdnl
primary10pripryprprim
broadcast10brdbrdcstbcst
Table 2. Ad Hoc abbreviations and their candidate definitions.
Table 2. Ad Hoc abbreviations and their candidate definitions.
Candidate Definition 1Candidate Definition 2Candidate Definition 3Candidate Definition 4Candidate Definition 5Candidate 5
bgbagbegarbeingbackgroundbigbgr
bkgdbackgatedbackgrindbackgroundbacksighted
bkgndbackgrindbackgroundbackgammoned
bkblinkbreakbackblackbackground
bgrbegarbaggerbiggerbackgroundburger
Table 3. Candidate definition statistics.
Table 3. Candidate definition statistics.
Definition 1Definition 2Definition 3Definition 4Definition 5# Candidate Words)Letter RatioLetters RemovedVowels FoundVowls KeptVowels Removed
bgbagbegarbeingbackgroundbig1000+0.28303
bkgdbackgatedbackgrindbackgroundbacksighted 530.46303
bkgndbackgrindbackgroundbackgammoned 170.64303
bkblinkbreakbackblackbackground1000+0.28303
bgrbegarbaggerbiggerbackgroundburger1000+0.37303
addladdleadducibleadditionaladdressibleadjudgeable8600.46404
addtladditionaladjudicationaladductivelyautodidacticallyaddictionologists470.55404
addnladdentaladditionalantroduodenaladdictionologist 650.55404
adjadjunctadjutantadjacentadjustadjectives5820.53101
adadvanceabideaddedadmitahead1000+0.3333333332303
allocallocateallowanceallotheticallocriteallogenic1000+0.55413
modsmodishmodulesmodificationmodelssmicropods1000+0.3076923089615
priperilprideprimaprimarypraise1000+0.4285714294211
pryproxyprayerpoetrypurifyprimary1000+0.4285714294202
prprompropproxyprimaryprofessor1000+0.2857142865202
primprimalpluralismprelimitsprimaryperimeter1000+0.5714285713211
brdbroadcastbirdsbreedbroiledbromide1000+0.3333333336303
brdcstbreadcrustbroadcast 460.6666666673303
bcstbackseatbackstepbuckshotbiochemistbroadcast1000+0.4444444445303
Table 4. Probability of selecting definition.
Table 4. Probability of selecting definition.
Target DefinitionCandidate DefinitionCategory NumberTarget ProbabilityCandidate Probability
adjustadjacent20.008540.02135
adjustadjacent140.003110.00623
adjustadjacent160.001460.00439
advanceadded130.025370.00260
modificationmodules140.003570.00936
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Choi, S.; Taghva, K. Findings on Ad Hoc Contractions. Information 2023, 14, 391. https://doi.org/10.3390/info14070391

AMA Style

Choi S, Taghva K. Findings on Ad Hoc Contractions. Information. 2023; 14(7):391. https://doi.org/10.3390/info14070391

Chicago/Turabian Style

Choi, Sing, and Kazem Taghva. 2023. "Findings on Ad Hoc Contractions" Information 14, no. 7: 391. https://doi.org/10.3390/info14070391

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop