Search results for "File system"

showing 10 items of 31 documents

Spontaneous quantity discrimination of artificial flowers by foraging honeybees

2020

ABSTRACTMany animals need to process numerical and quantity information in order to survive. Spontaneous quantity discrimination allows differentiation between two or more quantities without reinforcement or prior training on any numerical task. It is useful for assessing food resources, aggressive interactions, predator avoidance and prey choice. Honeybees have previously demonstrated landmark counting, quantity matching, use of numerical rules, quantity discrimination and arithmetic, but have not been tested for spontaneous quantity discrimination. In bees, spontaneous quantity discrimination could be useful when assessing the quantity of flowers available in a patch and thus maximizing f…

0106 biological sciencesPhysiology[SDV]Life Sciences [q-bio]ForagingSubitizingFlowersNumericAquatic Science010603 evolutionary biology01 natural sciencesPredation03 medical and health sciences0302 clinical medicineStatisticsApproximate number systemApproximate number systemAnimalsPredator avoidanceMolecular BiologyRatioEcology Evolution Behavior and SystematicsMathematicsArtificial flowerBees[SDV] Life Sciences [q-bio]Food resourcesInsect ScienceObject file systemAnimal Science and ZoologyApis mellifera030217 neurology & neurosurgery
researchProduct

Moderated Redactable Blockchains: A Definitional Framework with an Efficient Construct

2020

Blockchain is a multiparty protocol to reach agreement on the order of events, and to record them consistently and immutably without centralized trust. In some cases, however, the blockchain can benefit from some controlled mutability. Examples include removing private information or unlawful content, and correcting protocol vulnerabilities which would otherwise require a hard fork. Two approaches to control the mutability are: moderation, where one or more designated administrators can use their private keys to approve a redaction, and voting, where miners can vote to endorse a suggested redaction. In this paper, we first present several attacks against existing redactable blockchain solut…

020203 distributed computingComputer sciencemedia_common.quotation_subject02 engineering and technologyConstruct (python library)RedactionComputer securitycomputer.software_genreDigital signatureOrder (exchange)020204 information systemsVoting0202 electrical engineering electronic engineering information engineeringFork (file system)Protocol (object-oriented programming)computerPrivate information retrievalmedia_common
researchProduct

FASTA/Q data compressors for MapReduce-Hadoop genomics: space and time savings made easy

2021

Abstract Background Storage of genomic data is a major cost for the Life Sciences, effectively addressed via specialized data compression methods. For the same reasons of abundance in data production, the use of Big Data technologies is seen as the future for genomic data storage and processing, with MapReduce-Hadoop as leaders. Somewhat surprisingly, none of the specialized FASTA/Q compressors is available within Hadoop. Indeed, their deployment there is not exactly immediate. Such a State of the Art is problematic. Results We provide major advances in two different directions. Methodologically, we propose two general methods, with the corresponding software, that make very easy to deploy …

Big DataFASTQ formatComputer scienceBig data02 engineering and technologycomputer.software_genrelcsh:Computer applications to medicine. Medical informaticsBiochemistry03 medical and health sciencesSoftwareStructural BiologySpark (mathematics)0202 electrical engineering electronic engineering information engineeringData_FILESMapReduceMapReduce; hadoop; sequence analysis; data compressionMolecular Biologylcsh:QH301-705.5030304 developmental biologyFile system0303 health sciencesSettore INF/01 - InformaticaDatabasebusiness.industryMethodology ArticleApplied MathematicsSequence analysisGenomicsData compression; Hadoop; MapReduce; Sequence analysis; Algorithms; Big Data; Data Compression; Genomics; SoftwareComputer Science Applicationslcsh:Biology (General)Software deploymentHadoopData compressionlcsh:R858-859.7020201 artificial intelligence & image processingState (computer science)businesscomputerAlgorithmsSoftwareData compressionBMC Bioinformatics
researchProduct

AIOC2: A deep Q-learning approach to autonomic I/O congestion control in Lustre

2021

Abstract In high performance computing systems, I/O congestion is a common problem in large-scale distributed file systems. However, the current implementation mainly requires administrator to manually design low-level implementation and optimization, we proposes an adaptive I/O congestion control framework, named AIOC 2 , which can not only adaptively tune the I/O congestion control parameters, but also exploit the deep Q-learning method to start the training parameters and optimize the tuning for different types of workloads from the server and the client at the same time. AIOC 2 combines the feedback-based dynamic I/O congestion control and deep Q-learning parameter tuning technology to …

ExploitComputer Networks and CommunicationsComputer sciencebusiness.industryQ-learningInterference (wave propagation)SupercomputerComputer Graphics and Computer-Aided DesignTheoretical Computer ScienceNetwork congestionArtificial IntelligenceHardware and ArchitectureEmbedded systemLustre (file system)Latency (engineering)businessThroughput (business)SoftwareParallel Computing
researchProduct

GekkoFS - A Temporary Distributed File System for HPC Applications

2018

We present GekkoFS, a temporary, highly-scalable burst buffer file system which has been specifically optimized for new access patterns of data-intensive High-Performance Computing (HPC) applications. The file system provides relaxed POSIX semantics, only offering features which are actually required by most (not all) applications. It is able to provide scalable I/O performance and reaches millions of metadata operations already for a small number of nodes, significantly outperforming the capabilities of general-purpose parallel file systems. The work has been funded by the German Research Foundation (DFG) through the ADA-FS project as part of the Priority Programme 1648. It is also support…

File system020203 distributed computingBurst buffersParallel processing (Electronic computers)Computer scienceProcessament en paral·lel (Ordinadors)020207 software engineering02 engineering and technologyBuffer storage (Computer science)computer.software_genreData structureDistributed file systemsMetadataParallel processing (DSP implementation)POSIXServerScalabilityHPC0202 electrical engineering electronic engineering information engineeringOperating systemHigh performance computingDistributed File System:Informàtica::Arquitectura de computadors::Arquitectures paral·leles [Àrees temàtiques de la UPC]computerCàlcul intensiu (Informàtica)2018 IEEE International Conference on Cluster Computing (CLUSTER)
researchProduct

Sorted deduplication: How to process thousands of backup streams

2016

The requirements of deduplication systems have changed in the last years. Early deduplication systems had to process dozens to hundreds of backup streams at the same time while today they are able to process hundreds to thousands of them. Traditional approaches rely on stream-locality, which supports parallelism, but which easily leads to many non-contiguous disk accesses, as each stream competes with all other streams for the available resources. This paper presents a new exact deduplication approach designed for processing thousands of backup streams at the same time on the same fingerprint index. The underlying approach destroys the traditionally exploited temporal chunk locality and cre…

File system020203 distributed computingComputer scienceData domainFingerprint (computing)Search engine indexingSorting020206 networking & telecommunications02 engineering and technologyParallel computingcomputer.software_genreBackupServerData_FILES0202 electrical engineering electronic engineering information engineeringData deduplicationcomputer2016 32nd Symposium on Mass Storage Systems and Technologies (MSST)
researchProduct

DelveFS - An Event-Driven Semantic File System for Object Stores

2020

Data-driven applications are becoming increasingly important in numerous industrial and scientific fields, growing the need for scalable data storage, such as object storage. Yet, many data-driven applications cannot use object interfaces directly and often have to rely on third-party file system connectors that support only a basic representation of objects as files in a flat namespace. With sometimes millions of objects per bucket, this simple organization is insufficient for users and applications who are usually only interested in a small subset of objects. These huge buckets are not only lacking basic semantic properties and structure, but they are also challenging to manage from a tec…

File system020203 distributed computingDatabaseEvent (computing)business.industryComputer scienceRepresentation (systemics)020206 networking & telecommunications02 engineering and technologyDirectorycomputer.software_genreObject (computer science)Object storageComputer data storageScalability0202 electrical engineering electronic engineering information engineeringbusinesscomputer2020 IEEE International Conference on Cluster Computing (CLUSTER)
researchProduct

LPCC

2019

Most high-performance computing (HPC) clusters use a global parallel file system to enable high data throughput. The parallel file system is typically centralized and its storage media are physically separated from the compute cluster. Compute nodes as clients of the parallel file system are often additionally equipped with SSDs. The node internal storage media are rarely well-integrated into the I/O and compute workflows. How to make full and flexible use of these storage media is therefore a valuable research question. In this paper, we propose a hierarchical Persistent Client Caching (LPCC) mechanism for the Lustre file system. LPCC provides two modes: RW-PCC builds a read-write cache on…

File systemComputer scienceComputer clusterHierarchical storage management0202 electrical engineering electronic engineering information engineeringOperating system020206 networking & telecommunications020207 software engineeringLustre (file system)02 engineering and technologyCachecomputer.software_genrecomputerProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis
researchProduct

A configurable rule based classful token bucket filter network request scheduler for the lustre file system

2017

HPC file systems today work in a best-effort manner where individual applications can flood the file system with requests, effectively leading to a denial of service for all other tasks. This paper presents a classful Token Bucket Filter (TBF) policy for the Lustre file system. The TBF enforces Remote Procedure Call (RPC) rate limitations based on (potentially complex) Quality of Service (QoS) rules. The QoS rules are enforced in Lustre's Object Storage Servers, where each request is assigned to an automatically created QoS class.The proposed QoS implementation for Lustre enables various features for each class including the support for high-priority and real-time requests even under heavy …

File systemComputer scienceQuality of service020206 networking & telecommunications020207 software engineeringDenial-of-service attackRule-based system02 engineering and technologycomputer.software_genreObject storageRemote procedure callServer0202 electrical engineering electronic engineering information engineeringOperating systemLustre (file system)computerProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis
researchProduct

Challenges and Solutions for Tracing Storage Systems

2018

IBM Spectrum Scale’s parallel file system General Parallel File System (GPFS) has a 20-year development history with over 100 contributing developers. Its ability to support strict POSIX semantics across more than 10K clients leads to a complex design with intricate interactions between the cluster nodes. Tracing has proven to be a vital tool to understand the behavior and the anomalies of such a complex software product. However, the necessary trace information is often buried in hundreds of gigabytes of by-product trace records. Further, the overhead of tracing can significantly impact running applications and file system performance, limiting the use of tracing in a production system. In…

File systemComputer sciencebusiness.industryInterface (computing)Distributed computing020206 networking & telecommunications020207 software engineering02 engineering and technologyTracingcomputer.software_genreSoftwareHardware and ArchitecturePOSIXScalability0202 electrical engineering electronic engineering information engineeringOverhead (computing)businesscomputerTRACE (psycholinguistics)ACM Transactions on Storage
researchProduct