Introduction: This study analyzes the scientific production in business administration in scientific articles based on modeling partial least squares structural equations (Partial Least Squares Structural Equation Modeling PLS-SEM) in the 2011-2020 period. Methodology: The study is exploratory - descriptive and has three phases: a) Selection of keywords and search criteria; (b) Search and refinement of information; c) information analysis. A method of bibliometric review of the specific literature has been used based on the analysis of predefined indicators and completed with a qualitative content synthesis. Results: A total of 167 publications were analyzed, making correlations from the year, search criteria, authors, impact factor by quartile, and by citation variables. More outstanding scientific production comes from Scopus under the search criteria ((pls AND sem) OR "partial least squares") AND (business OR management), being the figure of 4,870 scientific articles, while Web of Science accumulates 3,946 articles Conclusion: There has been a progressive growth in scientific articles with the PLS-SEM technique from 2011 to 2020. Scopus, compared to WoS, presents a more significant number of scientific productions with this statistical approach. The authors who register scientific articles demonstrate a high H index; in addition, there is an important number of scientific articles with a PLS-SEM approach in universities in Malaysia that could be related to the expansion of higher education in that country, as well as in Singapore, Taiwan, and Indonesia. Finally, business administration, accounting, and economics are outstanding scientific production.
In recent years, with the rapid growth of Internet data, the number and types of scientific and technological resources are also rapidly expanding. However, the increase in the number and category of information data will also increase the cost of information acquisition. For technology-based enterprises or users, in addition to general papers, patents, etc., policies related to technology or the development of their industries should also belong to a type of scientific and technological resources. The cost and difficulty of acquiring users. Extracting valuable science and technology policy resources from a huge amount of data with mixed contents and providing accurate and fast retrieval will help to break down information barriers and reduce the cost of information acquisition, which has profound social significance and social utility. This article focuses on the difficulties and problems in the field of science and technology policy, and introduces related technologies and developments.
Readability assessment is the task of evaluating the reading difficulty of a given piece of text. Although research on computational approaches to readability assessment is now two decades old, there is not much work on synthesizing this research. This article is a brief survey of contemporary research on developing computational models for readability assessment. We identify the common approaches, discuss their shortcomings, and identify some challenges for the future. Where possible, we also connect computational research with insights from related work in other disciplines such as education and psychology.
Forensic firearms identification, the determination by a trained firearms examiner as to whether or not bullets or cartridges came from a common weapon, has long been a mainstay in the criminal courts. Reliability of forensic firearms identification has been challenged in the general scientific community, and, in response, several studies have been carried out aimed at showing that firearms examination is accurate, that is, has low error rates. Less studied has been the question of consistency, of. whether two examinations of the same bullets or cartridge cases come to the same conclusion, carried out by an examiner on separate occasions -- intrarater reliability or repeatability -- or by two examiners -- interrater reliability or reproducibility. One important study, described in a 2020 Report by the Ames Laboratory-USDOE to the Federal Bureau of Investigation, went beyond considerations of accuracy to investigate firearms examination repeatability and reproducibility. The Report's conclusions were paradoxical. The observed agreement of examiners with themselves or with other examiners appears mediocre. However, the study concluded repeatability and reproducibility are satisfactory, on grounds that the observed agreement exceeds a quantity called the expected agreement. We find that appropriately employing expected agreement as it was intended does not suggest satisfactory repeatability and reproducibility, but the opposite.
Microcosm, Hyper-G, and the Web were developed and released after 1989. There were strengths and weaknesses associate with each of these hypertext systems. The architectures of these systems were relatively different from one another. Standing above its competitors, the Web became the largest and most popular information system. This paper analyses the reasons for which the Web became the first successful hypermedia system by looking and evaluating the architecture of the Web, Hyper-G, and Microcosm systems. Three reasons will be given beyond this success with some lessons to learn. Currently, Semantic Web is a recent development of the Web to provide conceptual hypermedia. More importantly, study of the Web with its impact on technical, socio-cultural, and economical agendas is introduced as web science.
With the advent of open source software, a veritable treasure trove of previously proprietary software development data was made available. This opened the field of empirical software engineering research to anyone in academia. Data that is mined from software projects, however, requires extensive processing and needs to be handled with utmost care to ensure valid conclusions. Since the software development practices and tools have changed over two decades, we aim to understand the state-of-the-art research workflows and to highlight potential challenges. We employ a systematic literature review by sampling over one thousand papers from leading conferences and by analyzing the 286 most relevant papers from the perspective of data workflows, methodologies, reproducibility, and tools. We found that an important part of the research workflow involving dataset selection was particularly problematic, which raises questions about the generality of the results in existing literature. Furthermore, we found a considerable number of papers provide little or no reproducibility instructions -- a substantial deficiency for a data-intensive field. In fact, 33% of papers provide no information on how their data was retrieved. Based on these findings, we propose ways to address these shortcomings via existing tools and also provide recommendations to improve research workflows and the reproducibility of research.
The quest to understand consciousness, once the purview of philosophers and theologians, is now actively pursued by scientists of many stripes. We examine consciousness from the perspective of theoretical computer science (TCS), a branch of mathematics concerned with understanding the underlying principles of computation and complexity, including the implications and surprising consequences of resource limitations. In the spirit of Alan Turing's simple yet powerful definition of a computer, the Turing Machine (TM), and perspective of computational complexity theory, we formalize a modified version of the Global Workspace Theory (GWT) of consciousness originated by cognitive neuroscientist Bernard Baars and further developed by him, Stanislas Dehaene, Jean-Pierre Changeaux and others. We are not looking for a complex model of the brain nor of cognition, but for a simple computational model of (the admittedly complex concept of) consciousness. We do this by defining the Conscious Turing Machine (CTM), also called a conscious AI, and then we define consciousness and related notions in the CTM. While these are only mathematical (TCS) definitions, we suggest why the CTM has the feeling of consciousness. The TCS perspective provides a simple formal framework to employ tools from computational complexity theory and machine learning to help us understand consciousness and related concepts. Previously we explored high level explanations for the feelings of pain and pleasure in the CTM. Here we consider three examples related to vision (blindsight, inattentional blindness, and change blindness), followed by discussions of dreams, free will, and altered states of consciousness.
Present-day atomistic simulations generate long trajectories of ever more complex systems. Analyzing these data, discovering metastable states, and uncovering their nature is becoming increasingly challenging. In this paper, we first use the variational approach to conformation dynamics to discover the slowest dynamical modes of the simulations. This allows the different metastable states of the system to be located and organized hierarchically. The physical descriptors that characterize metastable states are discovered by means of a machine learning method. We show in the cases of two proteins, Chignolin and Bovine Pancreatic Trypsin Inhibitor, how such analysis can be effortlessly performed in a matter of seconds. Another strength of our approach is that it can be applied to the analysis of both unbiased and biased simulations.
One of the first steps in an academic career, and perhaps the pillar thereof, is completing a PhD under the supervision of a doctoral advisor. While prior work has examined the advisor-advisee relationship and its potential effects on the prospective academic success of the advisee, very little is known on the possibly continued relationship after the advisee has graduated. We harnessed three genealogical and scientometric datasets to identify 3 distinct groups of computer scientists: Highly independent, who cease collaborating with their advisors (almost) instantly upon graduation; Moderately independent, who (quickly) reduce the collaboration rate over ~5 years; and Weakly independent who continue collaborating with their advisors for at least 10 years post-graduation. We find that highly independent researchers are more academically successful than their peers in terms of H-index, i10-index and total number of citations throughout their careers. Moderately independent researchers perform, on average, better than weakly independent researchers, yet the differences are not found to be statistically significant. In addition, both highly and moderately independent researchers are found to have longer academic careers. Interestingly, weakly independent researchers tend to be supervised by more academically successful advisors.
Predictions obtained by, e.g., artificial neural networks have a high accuracy but humans often perceive the models as black boxes. Insights about the decision making are mostly opaque for humans. Particularly understanding the decision making in highly sensitive areas such as healthcare or fifinance, is of paramount importance. The decision-making behind the black boxes requires it to be more transparent, accountable, and understandable for humans. This survey paper provides essential definitions, an overview of the different principles and methodologies of explainable Supervised Machine Learning (SML). We conduct a state-of-the-art survey that reviews past and recent explainable SML approaches and classifies them according to the introduced definitions. Finally, we illustrate principles by means of an explanatory case study and discuss important future directions.
The notion of uncertainty is of major importance in machine learning and constitutes a key element of machine learning methodology. In line with the statistical tradition, uncertainty has long been perceived as almost synonymous with standard probability and probabilistic predictions. Yet, due to the steadily increasing relevance of machine learning for practical applications and related issues such as safety requirements, new problems and challenges have recently been identified by machine learning scholars, and these problems may call for new methodological developments. In particular, this includes the importance of distinguishing between (at least) two different types of uncertainty, often refereed to as aleatoric and epistemic. In this paper, we provide an introduction to the topic of uncertainty in machine learning as well as an overview of hitherto attempts at handling uncertainty in general and formalizing this distinction in particular.