EBookClubs

Read Books & Download eBooks Full Online

EBookClubs

Read Books & Download eBooks Full Online

Book Assessing the Accuracy of Google Translate to Allow Data Extraction from Trials Published in Non English Languages

Download or read book Assessing the Accuracy of Google Translate to Allow Data Extraction from Trials Published in Non English Languages written by U. S. Department of Health and Human Services and published by Createspace Independent Pub. This book was released on 2013-03-21 with total page 44 pages. Available in PDF, EPUB and Kindle. Book excerpt: One of the strengths of systematic reviews is that they aim to include all relevant evidence. However, study eligibility is often restricted to the English language for practical reasons. Google Translate, a free Web-based resource for translation, has recently become available. However, it is unclear whether its translation accuracy is sufficient for systematic reviews. An earlier pilot study provided some evidence that data extraction from translated articles may be adequate but varies by language. To address several limitations of the pilot study, four collaborating Evidence-based Practice Centers conducted a more rigorous analysis of translations of articles from five languages. Systematic reviews conducted by the Agency for Healthcare Research and Quality (AHRQ) Evidence-based Practice Centers (EPCs) most commonly restrict literature searches to English language publications. In a sample of 10 recent Evidence Reports (numbers 189-198), 8 were restricted to English-language publications. One report included studies in languages for which the EPC had “available fluency” and only one reported not restricting by language. Among 28 other recent Comparative Effectiveness Reviews (CERs) with final or draft documents downloadable from the AHRQ Web site, 20 were restricted to English-language publications. Four explicitly did not impose any language restriction. Two did not report language restriction in their methods chapter and included one study each in Dutch and German. One placed no language restriction on comparative studies but included only English-language cohort studies. One included German- and French-language studies for nonoperative interventions (which were sparse), but only English-language publications for operative treatments “due to lack of translation resources.” Three of the CERs wrote that the language restriction was due to lack of resources or prohibitive translation costs, despite the recognition in one CER “that requiring studies to be published in English could lead to bias.” The current study was designed to form a collaboration of EPCs to better analyze the accuracy of the freely available, online, translation tool—Google Translate—for the purposes of data extraction of articles in selected non-English languages. The collaboration allowed for double data extraction and a better consensus determination of the important extraction items to assess; we also implemented an improved analytic technique. The research had the following aims: 1. Compare data extraction of trials done on original-language articles by native speakers with data extraction done on articles translated to English by Google Translate. 2. Track and enumerate the time and resources used for article translation and the extra time and resources required for data extraction related to use of translated articles.

Book Assessing the Accuracy of Google Translate to Allow Data Extraction from Trials Published in Non English Languages

Download or read book Assessing the Accuracy of Google Translate to Allow Data Extraction from Trials Published in Non English Languages written by Ethan Balk and published by . This book was released on 2013 with total page pages. Available in PDF, EPUB and Kindle. Book excerpt: BACKGROUND: One of the strengths of systematic reviews is that they aim to include all relevant evidence. However, study eligibility is often restricted to the English language for practical reasons. Google Translate, a free Web-based resource for translation, has recently become available. However, it is unclear whether its translation accuracy is sufficient for systematic reviews. An earlier pilot study provided some evidence that data extraction from translated articles may be adequate but varies by language. To address several limitations of the pilot study, four collaborating Evidence-based Practice Centers conducted a more rigorous analysis of translations of articles from five languages. METHODS: We included 10 randomized controlled trials in 5 languages (Chinese, French, German, Japanese, and Spanish). Eligible studies were trials that reported per-treatment group results data. Each article was translated into English using Google Translate. The time required to translate each study was tracked. The original language versions of the articles were double data extracted by fluent speakers and reconciled. Each English-translated article was extracted by two of eight researchers who did not speak the given language. These 8 researchers also each extracted 10 English-language trials to serve as a control. Data extracted included: eligibility criteria, study design features, outcomes reported, intervention and outcome descriptions, and results data for one continuous and/or one categorical outcome. We used a generalized linear mixed model to examine whether the probability of correctly extracting an item from a translated article is related to the language of original publication. The model used each extractor's accuracy in extracting the English language trials to control for reviewer effects. RESULTS: The length of time required to translate articles ranged from 5 minutes to about 1 hour for almost all articles, with an average of about 30 minutes. Extractors estimated that most Spanish articles required less than 5 additional minutes to extract because of translation issues, but about two-thirds of other language articles required between 6 and 30 additional minutes for extraction. Analyses of the adjusted percentage of correct extractions across items and languages and of the adjusted odds ratio of correct extractions compared with English revealed that in general, across languages the likelihood of correct extractions was greater for study design and intervention domain items than for outcome descriptions and, particularly, study results. Translated Chinese articles yielded the highest percentage of items (22 percent) that were incorrectly extracted more than half the time (but also the largest percentage of items, 41 percent, that were extracted correctly more than 98 percent of the time). Relative to English, extractions of translated Spanish articles were most accurate compared with other translated languages. CONCLUSION: Translation generally required few resources. Across all languages, data extraction from translated articles was less accurate than from English language articles, particularly and importantly for results data. Extraction was most accurate from translated Spanish articles and least accurate from translated Chinese articles. Use of Google Translate has the potential of being an approach to reduce language bias; however, reviewers may need to be more cautious about using data from these translated articles. There remains a tradeoff between completeness of systematic reviews (including all available studies) and risk of error (due to poor translation).

Book Enhancements and Limitations to ICT Based Informal Language Learning  Emerging Research and Opportunities

Download or read book Enhancements and Limitations to ICT Based Informal Language Learning Emerging Research and Opportunities written by Ahmed, Rashad and published by IGI Global. This book was released on 2020-01-03 with total page 201 pages. Available in PDF, EPUB and Kindle. Book excerpt: In today’s modernized world, digital technology has taken the forefront in all aspects of society, including education. Students have access to numerous electronic devices, which has made online learning materials highly accessible. These technological impacts have blurred the distinction between formal and informal language learning methods. Informally learned English has lost proficiency when assessing student performance. Sizable research is necessary to study and understand the informal methods of language learning using technology. Enhancements and Limitations to ICT-Based Informal Language Learning: Emerging Research and Opportunities is a pivotal reference source that provides vital research on the implementation of technological opportunities within informal language teaching methods along with the drawbacks that limit its efficiency. While highlighting topics such as acculturation, student perception, and autonomous applications, this publication explores how learners perform ICT-based activities beyond the classroom and assesses the linguistic gains generated by informal ICT uses. This book is ideally designed for teachers, IT consultants, educational software developers, researchers, policymakers, and academic professionals seeking current research on technological techniques within second language learning and teaching.

Book Principles and Practice of Systematic Reviews and Meta Analysis

Download or read book Principles and Practice of Systematic Reviews and Meta Analysis written by Sanjay Patole and published by Springer Nature. This book was released on 2021-06-26 with total page 189 pages. Available in PDF, EPUB and Kindle. Book excerpt: Evidence based medicine is at the core of modern medicine. It involves the integration of individual clinical expertise with the best available clinical evidence from systematic research and patient's values and expectations. Systematic reviews offer a summary of the best available evidence. They are the most reliable and comprehensive statement about what works. Written by clinical academics from Australia, UK, USA, and Switzerland, this contributed volume introduces the readers to the principles and practice of systematic reviews and meta-analysis. It covers the various steps involved in systematic reviews including development of a focused question and the strategy for conducting a comprehensive literature search, identifying studies addressing the underlying question, assessment of heterogeneity and the risk of bias in the included studies, data extraction, and the approach to meta-analysis. Crucial issues such as selecting the model for meta-analysis, generating and interpreting forest plots, assessing the risk of publication bias, cautions in the interpretation of subgroup and sensitivity analyses, rating certainty of the evidence using GRADE guideline, and standardized reporting of meta-analysis (PRISMA) are covered in detail. Every attempt is made to keep the narrative simple and clear. Mathematical formulae are avoided as much as possible. While the focus of this book is on systematic reviews and meta-analyses of randomised controlled trials (RCTs), the gold standard of clinical research, the essentials of systematic reviews of non-RCTs, diagnostic test accuracy studies, animal studies, individual participant data meta-analysis, and network meta-analysis are also covered. Readers from all faculties of medicine will enjoy this comprehensive and reader friendly book to understand the principles and practice of systematic reviews and meta-analysis for guiding their clinical practice and research.

Book Corpus based Approaches to Grammar  Media and Health Discourses

Download or read book Corpus based Approaches to Grammar Media and Health Discourses written by Bingjun Yang and published by Springer Nature. This book was released on 2020-06-23 with total page 395 pages. Available in PDF, EPUB and Kindle. Book excerpt: This edited volume gathers corpus-based studies on topics including English grammar and discourses on media and health, mainly from a systemic functional linguistics (SFL) perspective, in order to reveal the potential of SFL, which has been emphasized by Halliday. Various other perspectives, such as philosophy, statistics, genre studies, etc. are also included to promote SFL’s potential interaction with other theories. Though they employ a diverse range of theoretical perspectives, all the chapters focus on exploring language in use with the corpus method. The studies collected here are all original, unpublished research articles that address significant questions, deepen readers’ understanding of SFL, and promote its potential interaction with other theories. In addition, they demonstrate the great potential that SFL holds for solving language-related questions in a variety of discourses.

Book Recent Developments in Individual and Organizational Adoption of ICTs

Download or read book Recent Developments in Individual and Organizational Adoption of ICTs written by Yildiz, Orkun and published by IGI Global. This book was released on 2020-08-28 with total page 333 pages. Available in PDF, EPUB and Kindle. Book excerpt: In recent years, information and communication technologies (ICTs) have gained significant importance and become vital to the operations of both organizations and individuals. However, there are numerous factors that have affected the adoption of ICTs including access and accessibility barriers, political participation, and social empowerment. This has attracted the attention of researchers who are interested in understanding the socioeconomic influences of ICT adoption and how these technologies impact the infrastructure of modern organizational activities. Recent Developments in Individual and Organizational Adoption of ICTs is a collection of innovative research on the methods of organizational and infrastructural advancement through the application of information and communication technologies. While highlighting topics including internet banking, supply chain management, and e-government services, this book is ideally designed for managers, researchers, policymakers, politicians, business practitioners, educators, decision scientists, strategists, and students seeking current research on the socioeconomic impact of ICT adoption.

Book Introduction to Google Translate

Download or read book Introduction to Google Translate written by Gilad James, PhD and published by Gilad James Mystery School. This book was released on with total page 90 pages. Available in PDF, EPUB and Kindle. Book excerpt: Google Translate is a multilingual translation service provided by Google. It allows users to translate words, phrases, and entire documents between multiple languages. The service was launched in April 2006 and has since been constantly updated to provide more accurate translations. Initially offering translations in only two languages, Google Translate now supports over 100 languages. The translation process works by analyzing the text or document input by the user, breaking it up into smaller segments, and then using statistical algorithms to match these segments with translations from its database. Google Translate has been a helpful tool for people to communicate across different languages, whether it be for business or personal use. However, it must be noted that automated translations often carry a high risk of inaccuracies due to the complexities inherent in language and the nuances of different cultures and contexts. It is always recommended to use translations as a starting point, and then have a native speaker review and refine the language to ensure accuracy.

Book Information Extraction to Facilitate Translation of Natural Language Legislation

Download or read book Information Extraction to Facilitate Translation of Natural Language Legislation written by Samuel Siyue Wang and published by . This book was released on 2011 with total page 66 pages. Available in PDF, EPUB and Kindle. Book excerpt: There is a large body of existing legislation and policies that govern how government organizations and corporations can share information. Since these rules are generally expressed in natural language, it is difficult and labor intensive to verify whether or not data sharing events are compliant with the relevant policies. This work aims to develop a natural language processing framework that automates significant portions of this translation process, so legal policies are more accessible to existing automated reasoning systems. Even though these laws are expressed in natural language, for this very specific domain, only a handful of sentence structures are actually used to convey logic. This structure can be exploited so that the program can automatically detect who the actor, action, object, and conditions are for each rule. In addition, once the structure of a rule is identified, similar rules can be presented to the user. If integrated into an authoring environment, this will allow the user to reuse previously translated rules as templates to translate novel rules more easily, independent of the target language for translation. A body of 315 real-world rules from 12 legal sources was collected and annotated for this project. Cross-validation experiments were conducted on this annotated data set, and the developed system was successful in identifying the underlying rule structure 43% of the time, and annotating the underlying tokens with recall of .66 and precision of .66. In addition, for 70% of the rules in each test set, the underlying rule structure had been seen in the training set. This suggests that the hypothesis that rules can only be expressed in a limited number of ways is probable.

Book How Google Translate and Bing Translator Cope with Grammar

Download or read book How Google Translate and Bing Translator Cope with Grammar written by Xiang Weng and published by . This book was released on 2011 with total page pages. Available in PDF, EPUB and Kindle. Book excerpt:

Book The Analysis of Google Translate as a Machine Translation Tool

Download or read book The Analysis of Google Translate as a Machine Translation Tool written by Špela Štular and published by . This book was released on 2018 with total page 65 pages. Available in PDF, EPUB and Kindle. Book excerpt: The MA thesis analyses the success of Google Translate as a machine translation (henceforth MT) tool. For the better understanding of the translation process, I describe some general definitions of and approaches to translation. The thesis discusses machine translation and gives a thorough overview of its development and its positive and negative aspects. In my MA thesis, I present different MT tools, those which are available for personal computers, and those that are available online. Furthermore, I focus the usefulness of such tools and the strategies of translations they employ. Finally, I describe the currently most successful MT tool, Google Translate. In the research part of the thesis, I compare translations of three types of texts made by Google. In my MA thesis, I include three different thematic fields of texts: tourism, since it is a topic with quite general language that is used every day, economics with its specific terminology and a literary text with some phraseology. Lastly, I analyse the mistakes with the aim to point out the reasons for their appearance.

Book Machine Translation of User generated Content

Download or read book Machine Translation of User generated Content written by Pintu Lohar and published by . This book was released on 2020 with total page 0 pages. Available in PDF, EPUB and Kindle. Book excerpt: The world of social media has undergone huge evolution during the last few years. With the spread of social media and online forums, individual users actively participate in the generation of online content in different languages from all over the world. Sharing of online content has become much easier than before with the advent of popular websites such as Twitter, Facebook etc. Such content is referred to as 'User-Generated Content' (UGC). Some examples of UGC are user reviews, customer feedback, tweets etc. In general, UGC is informal and noisy in terms of linguistic norms. Such noise does not create significant problems for human to understand the content, but it can pose challenges for several natural language processing applications such as parsing, sentiment analysis, machine translation (MT), etc. An additional challenge for MT is sparseness of bilingual (translated) parallel UGC corpora. In this research, we explore the general issues in MT of UGC and set some research goals from our findings. One of our main goals is to exploit comparable corpora in order to extract parallel or semantically similar sentences. To accomplish this task, we design a document alignment system to extract semantically similar bilingual document pairs using the bilingual comparable corpora. We then apply strategies to extract parallel or semantically similar sentences from comparable corpora by transforming the document alignment system into a sentence alignment system. We seek to improve the quality of parallel data extraction for UGC translation and assemble the extracted data with the existing human translated resources. Another objective of this research is to demonstrate the usefulness of MT-based sentiment analysis. However, when using openly available systems such as Google Translate, the translation process may alter the sentiment in the target language. To cope with this phenomenon, we instead build fine-grained sentiment translation models that focus on sentiment preservation in the target language during translation.

Book Measuring Translation Quality by Testing English Speakers with a New Defense Language Proficiency Test for Arabic

Download or read book Measuring Translation Quality by Testing English Speakers with a New Defense Language Proficiency Test for Arabic written by and published by . This book was released on 2005 with total page 7 pages. Available in PDF, EPUB and Kindle. Book excerpt: We present results from an experiment in which educated English-native speakers answered questions from a machine translated version of a standardized Arabic language test. We compare the machine translation (MT) results with professional reference translations as a baseline for the purpose of determining the level of Arabic reading comprehension that current machine translation technology enables an English speaker to achieve. Furthermore, we explore the relationship between the current, broadly accepted automatic measures of performance for machine translation and the Defense Language Proficiency Test, a broadly accepted measure of effectiveness for evaluating foreign language proficiency. In doing so, we intend to help translate MT system performance into terms that are meaningful for satisfying Government foreign language processing requirements. The results of this experiment suggest that machine translation may enable Interagency Language Roundtable Level 2 performance, but is not yet adequate to achieve ILR Level 3. Our results are based on 69 human subjects reading 68 documents and answering 173 questions, giving a total of 4,692 timed document trials and 7,950 question trials. We propose Level 3 as a reasonable near-term target for machine translation research and development.

Book Creating a Strong Statistical Machine Translation System by Combining Different Decoders

Download or read book Creating a Strong Statistical Machine Translation System by Combining Different Decoders written by Ayah ElMaghraby and published by . This book was released on 2018 with total page 148 pages. Available in PDF, EPUB and Kindle. Book excerpt: Abstract: Machine translation is a very important field in Natural Language Processing. The need for machine translation arises due to the increasing amount of data available online. Most of our data now is digital and this is expected to increase over time. Since human manual translation takes a lot of time and effort, machine translation is needed to cover all of the languages available. A lot of research has been done to make machine translation faster and more reliable between different language pairs. Machine translation is now being coupled with deep learning and neural networks. New topics in machine translation are being studied and tested like applying neural machine translation as a replacement to the classical statistical machine translation. In this thesis, we also study the effect of data-preprocessing and decoder type on translation output. We then demonstrate two ways to enhance translation from English to Arabic. The first approach uses a two-decoder system; the first decoder translates from English to Arabic and the second is a post-processing decoder that retranslates the first Arabic output to Arabic again to fix some of the translation errors. We then study the results of different kinds of decoders and their contributions to the test set. The results of this study lead to the second approach which combines different decoders to create a stronger one. The second approach uses a classifier to categorize the English sentences based on their structure. The output of the classifier is the decoder that is suited best to translate the English sentence. Both approaches increased the BLEU score albeit with different ranges. The classifier showed an increase of ~0.1 BLEU points while the post-processing decoder showed an increase of between ~0.3~11 BLEU points on two different test sets. Eventually we compare our results to Google translate to know how well we are doing in comparison to a well-known translator. Our best translation machine system scored 5 absolute points compared to Google translate in ISI corpus test set and we were 9 absolute points lower in the case of the UN corpus test set..

Book Measuring Acceptability of Machine Translated Enterprise Content

Download or read book Measuring Acceptability of Machine Translated Enterprise Content written by Sheila Castilho Monteiro de Sousa and published by . This book was released on 2016 with total page 0 pages. Available in PDF, EPUB and Kindle. Book excerpt: This research measures end-user acceptability of machine-translated enterprise content. In cooperation with industry partners, the acceptability of machine translated, post-edited and human translated texts, as well as source text were measured using a user-centred translation approach (Suojanen, Koskinen and Tuominen 2015). The source language was English and the target languages German, Japanese and Simplified Chinese. Even though translation quality assessment (TQA) is a key topic in the translation field, academia and industry greatly differ on how to measure quality. While academia is mostly concerned with the theory of translation quality, TQA in the industry is mostly performed by making use of arbitrary error typology models where "one size fits all". Both academia and industry greatly disregard the end user of those translations when assessing the translation quality and so, the acceptability of translated and un-translated content goes largely unmeasured. Measuring acceptability of translated text is important because it allows one to identify what impact the translation might have on the end user - the final readers of the translation. Different stakeholders will have different acceptability thresholds for different languages and content types; some will want high quality translation, others may make do with faster turnaround, lower quality, or may even prefer non-translated content compared with raw MT. Acceptability is defined as usability, quality and satisfaction. Usability, in turn, is defined as effectiveness, efficiency in a specified context of use (ISO 2002) and is measured via tasks recorded using an eye tracker. Quality is evaluated via a TQA questionnaire answered by professional translators, and the source content is also evaluated via metrics such as readability and syntactic complexity. Satisfaction is measured via three different approaches: web survey, post-task questionnaire, and translators' ranking. By measuring the acceptability of different post-editing levels for three target languages as well as the source content, this study aims to understand the different thresholds users may have regarding their tolerance to translation quality, taking into consideration the content type and language. Results show that the implementation of light post-editing directly and positively influences acceptability for German and Simplified Chinese languages, more so than for the Japanese language and, moreover, the findings of this research show that different languages have different thresholds for translation quality.