Search results for "computer.software_genre"
showing 10 items of 3858 documents
Improved SOM Learning using Simulated Annealing
2007
Self-Organizing Map (SOM) algorithm has been extensively used for analysis and classification problems. For this kind of problems, datasets become more and more large and it is necessary to speed up the SOM learning. In this paper we present an application of the Simulated Annealing (SA) procedure to the SOM learning algorithm. The goal of the algorithm is to obtain fast learning and better performance in terms of matching of input data and regularity of the obtained map. An advantage of the proposed technique is that it preserves the simplicity of the basic algorithm. Several tests, carried out on different large datasets, demonstrate the effectiveness of the proposed algorithm in comparis…
Lone Star Stack: Architecture of a Disk-Based Archival System
2014
The need for huge storage systems rises with the ever growing creation of data. With growing capacities and shrinking prices, "write once read sometimes" workloads become more common. New data is constantly added, rarely updated or deleted, and every stored byte might be read at any time - a common pattern for digital archives or big data scenarios. We present the Lone Star Stack, a disk based archival storage system building block that is optimized for high reliability and energy efficiency. It provides a POSIX file system interface that uses flash based storage for write-offloading and metadata and the disk-based Lone Star RAID for user data storage. The RAID attempts to spin down disks a…
LoneStar RAID
2016
The need for huge storage archives rises with the ever growing creation of data. With today’s big data and data analytics applications, some of these huge archives become active in the sense that all stored data can be accessed at any time. Running and evolving these archives is a constant tradeoff between performance, capacity, and price. We present the LoneStar RAID, a disk-based storage architecture, which focuses on high reliability, low energy consumption, and cheap reads. It is designed for MAID systems with up to hundreds of disk drives per server and is optimized for “write once, read sometimes” workloads. We use dedicated data and parity disks, and export the data disks as individu…
VLPZ: The Vehicular Location Privacy Zone
2016
International audience; One of the key challenges in the success of vehicular ad hoc networks (VANETs) is to consider the location privacy of drivers. Although, the pseudonym changing approach is suggested by standardization development organizations such as IEEE and ETSI, the development of an effective pseudonym changing strategy is still an open issue. The existing pseudonym changing strategies are either not effective to protect against the pseudonyms linking attacks or can have a negative impact on the VANETs’ applications. To address these issues, this paper proposes a new pseudonym changing strategy based on the Vehicular Location Privacy Zone (VLPZ), which is a roadside infrastructu…
Building Digital Government by XML
2005
Continuing innovations in information and communication technologies offer powerful tools for building digital government but, at the same time, in many environments they have lead into a number of heterogeneous, expensive, and inconsistent solutions. XML offers a common metalanguage and terminology to develop means for system and data integration, and for gradual transfer to more consistent formats in information assets. The paper describes ways for the use of XML in public administration and gives examples of the use, particularly, in Finland. The paper introduces XML standardization levels and types in public administration. Experiences of the long-term standardization of the Finnish par…
The Shuffle Product: New Research Directions
2015
In this paper we survey some recent researches concerning the shuffle operation that arise both in Formal Languages and in Combinatorics on Words.
The Multivariate Individual Selection of Diagnostic Tests and the Reserved Diagnostic Statement: An Optimum Combination of Two New Methods for the Co…
1984
A combination of two new methods for the diagnostic procedure in computer-aided differential diagnosis is presented. It is constructed on the basis of new results of our own in the field of mathematical decision theory and is demonstrated by the differential diagnosis of congenital heart diseases by means of ECG features.
Taxonomy of stock market indices
2000
We investigate sets of financial non-redundant and nonsynchronously recorded time series. The sets are composed by a number of stock market indices located all over the world in five continents. By properly selecting the time horizon of returns and by using a reference currency we find a meaningful taxonomy. The detection of such a taxonomy proves that interpretable information can be stored in a set of nonsynchronously recorded time series.
CROSSMAPPER: estimating cross-mapping rates and optimizing experimental design in multi-species sequencing studies
2020
Motivation Numerous sequencing studies, including transcriptomics of host-pathogen systems, sequencing of hybrid genomes, xenografts, mixed species systems, metagenomics and meta-transcriptomics, involve samples containing genetic material from divergent organisms. A crucial step in these studies is identifying from which organism each sequencing read originated, and the experimental design should be directed to minimize biases caused by cross-mapping of reads to incorrect source genomes. Additionally, pooling of sufficiently different genetic material into a single sequencing library could significantly reduce experimental costs but requires careful planning and assessment of the impact of…
A Bayesian Sequential Look at u-Control Charts
2005
We extend the usual implementation of u-control charts (uCCs) in two ways. First, we overcome the restrictive (and often inadequate) assumptions of the Poisson model; next, we eliminate the need for the questionable base period by using a sequential procedure. We use empirical Bayes(EB) and Bayes methods and compare them with the traditional frequentist implementation. EB methods are somewhat easy to implement, and they deal nicely with extra-Poisson variability (and, at the same time, informally check the adequacy of the Poisson assumption). However, they still need the base period. The sequential, full Bayes approach, on the other hand, also avoids this drawback of traditional u-charts. T…