Search results for "natural language processing"
showing 10 items of 413 documents
Evidentials and Epistemic Modality
2018
Abstract This chapter deals with the relation between the notional domains of information source and epistemic modality. It surveys various approaches to this relation and the cross-linguistic patterns of the way in which linguistic units (of diverse formats) with evidential or epistemic meanings develop extensions whereby they encroach into each other’s domains. Meaning extensions in either direction can adequately be captured, and confusion between both domains can be avoided, only if in the analysis of the meaning of such units (a) an onomasiological and semasiological perspective and (b) a coded-inferred divide are distinguished. Thus, epistemic extensions often arise as Generalized Con…
A Conversational Agent Based on a Conceptual Interpretation of a Data Driven Semantic Space
2005
In this work we propose an interpretation of the LSA framework which leads to a data-driven “conceptual” space creation suitable for an “intuitive” conversational agent. The proposed approach allows overcoming the limitations of traditional, rule-based, chat-bots, leading to a more natural dialogue.
A word prediction methodology for automatic sentence completion
2015
Word prediction generally relies on n-grams occurrence statistics, which may have huge data storage requirements and does not take into account the general meaning of the text. We propose an alternative methodology, based on Latent Semantic Analysis, to address these issues. An asymmetric Word-Word frequency matrix is employed to achieve higher scalability with large training datasets than the classic Word-Document approach. We propose a function for scoring candidate terms for the missing word in a sentence. We show how this function approximates the probability of occurrence of a given candidate word. Experimental results show that the proposed approach outperforms non neural network lang…
Rigal — a programming language for compiler writing
2005
A new programming language for compiler writing is described. The main data structures are atoms, lists and trees. The control structures are based on advanced pattern matching. All phases of compilation, including parsing, optimization and code generation, can be programmed in this language in short and readable form. Sample compiler written in RIGAL is presented.
Southeast Asia as a Linguistic Area
2006
The languages belonging to the families of Mon-Khmer, Tai, Sinitic, Hmong-Mien, and Chamic (Malayo-Polynesian) spoken in mainland Southeast Asia share a considerable number of properties due to contact-induced convergence. General characteristics of the languages in this area are the lack of obligatory grammatical categories and the central role of pragmatics (indeterminateness), syllabic morphology, and relatively flexible correlations between word class and syntactic positions (versatility). The following properties will be described in some more detail: word order, numeral classifiers, cooccurrence patterns of the verb with coverbs, tense-aspect-modality markers and directional verbs, an…
A survey on emotion detection: A lexicon based backtracking approach for detecting emotion from Bengali text
2017
Emotion recognition ability has been introduced as a core component of emotional competence. Every emotion has different ways to be expressed such as text, speech, lyrics etc. This paper reflects the current experimental study and their outcomes on emotion detection from different textual data. In case of lexicon-based analysis, the position of emotional lexicons really varies the state of an emotion. In this empirical study, our focus was to find how people use the emotional keywords to express their emotions. We have presented an emotion detection model to extract emotion from Bengali text at the sentence level. In order to detect emotion from Bengali text, we have considered two basic em…
Stroke Cognitive Medical Assistant (StrokeCMA)
2018
Stroke is the number two killer after heart disease since it is responsible for almost 10% of all deaths worldwide. The main problem with a stroke is a significant delay in treatment that happened mainly due to inappropriate detection of stroke symptoms or inability of a person to perform further necessary actions, and might cause death, permanent disabilities, as well as more expensive treatment and rehabilitation. Nowadays assessment of a stroke is done by human, following widely adopted FAST approach of stroke assessment. Since a human factor become one of the causes of treatment delay, offered solution will try to minimize this factor. Artificial Intelligence, Cognitive Computing, Machi…
HOWERD: A Hidden Markov Model for Automatic OWL-ERD Alignment
2016
The HOWERD model for estimating the most likely alignment between an OWL ontology and an Entity Relation Diagram (ERD) is presented. Automatic alignment between relational schema and ontology represents a big challenge in Semantic Web research due to the different expressiveness of these representations. A relational schema is less expressive than the ontology; this is a non trivial problem when accessing data via an ontology and for ontology storing by means of a relational schema. Existent alignment methodologies fail in loosing some contents of the involved representations because the ontology captures more semantic information, and several elements are left unaligned. HOWERD relies on a…
Automated scoring of teachers’ pedagogical content knowledge : a comparison between human and machine scoring
2020
To validly assess teachers’ pedagogical content knowledge (PCK), performance-based tasks with open-response formats are required. Automated scoring is considered an appropriate approach to reduce the resource-intensity of human scoring and to achieve more consistent scoring results than human raters. The focus is on the comparability of human and automated scoring of PCK for economics teachers. The answers of (prospective) teachers (N = 852) to six open-response tasks from a standardized and validated test were scored by two trained human raters and the engine “Educational SCoRIng Toolkit” (ESCRITO). The average agreement between human and computer ratings, κw = 0.66, suggests a convergent …
VEBO: Validation of E-R diagrams through ontologies and WordNet
2012
In the semantic web vision, ontologies are building blocks for providing applications with a high level description of the operating environment in support of interoperability and semantic capabilities. The importance of ontologies in this respect is clearly stated in many works. Another crucial issue to increase the semantic aspect of web is to enrich the level of expressivity of database related data. Nowadays, databases are the primary source of information for dynamical web sites. The linguistic data used to build the database structure could be relevant for extracting meaningful information. In most cases, this type of information is not used for information retrieval. The work present…