Search results for "Compres"

showing 10 items of 1107 documents

Efficiency improvement of DC* through a Genetic Guidance

2017

DC∗ is a method for generating interpretable fuzzy information granules from pre-classified data. It is based on the subsequent application of LVQ1 for data compression and an ad-hoc procedure based on A∗ to represent data with the minimum number of fuzzy information granules satisfying some interpretability constraints. While being efficient in tackling several problems, the A∗ procedure included in DC∗ may happen to require a long computation time because the A∗ algorithm has exponential time complexity in the worst case. In this paper, we approach the problem of driving the search process of A∗ by suggesting a close-to-optimal solution that is produced through a Genetic Algorithm (GA). E…

Exponential complexity0209 industrial biotechnologyMathematical optimizationComputationProcess (computing)02 engineering and technologyFuzzy logic020901 industrial engineering & automationGenetic algorithm0202 electrical engineering electronic engineering information engineering020201 artificial intelligence & image processingAlgorithmMathematicsInterpretabilityData compression2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)
researchProduct

Bond length compressibility in hard ReB2 investigated by x-ray absorption under high pressure

2010

International audience; This work describes x-ray absorption measurements under high pressure in ReB2 , complemented by ab initio calculations. The EXAFS analysis yields the average Re–B bond compressibility, which turns out to be χReB = 5.6(9) × 10− 4 GPa−1 . Combining this information with previous x-ray diffraction experiments we have characterized the network of covalent bonds responsible for the rigidity of the structure. The main conclusion is that the compression is anisotropic and nonhomogeneous, reflecting bonding differences between Re–B and B–B bonds and also between nonequivalent Re–B bonds. The layer defined by boron atoms tends to become flatter under high pressure. As a conse…

Extended X-ray absorption fine structureChemistry02 engineering and technology021001 nanoscience & nanotechnologyCondensed Matter Physics01 natural sciencesMolecular physicsBond lengthCrystallographyChemical bondCovalent bondAb initio quantum chemistry methods[PHYS.COND.CM-GEN]Physics [physics]/Condensed Matter [cond-mat]/Other [cond-mat.other]0103 physical sciencesX-ray crystallographyCompressibilityGeneral Materials Science010306 general physics0210 nano-technologyStructural rigidity
researchProduct

On the Greedy Algorithm for the Shortest Common Superstring Problem with Reversals

2015

We study a variation of the classical Shortest Common Superstring (SCS) problem in which a shortest superstring of a finite set of strings $S$ is sought containing as a factor every string of $S$ or its reversal. We call this problem Shortest Common Superstring with Reversals (SCS-R). This problem has been introduced by Jiang et al., who designed a greedy-like algorithm with length approximation ratio $4$. In this paper, we show that a natural adaptation of the classical greedy algorithm for SCS has (optimal) compression ratio $\frac12$, i.e., the sum of the overlaps in the output string is at least half the sum of the overlaps in an optimal solution. We also provide a linear-time implement…

FOS: Computer and information sciences0102 computer and information sciences02 engineering and technologyInformation System01 natural sciencesString (physics)Theoretical Computer ScienceCombinatoricsHigh Energy Physics::TheoryAnalysis of algorithmGreedy algorithmComputer Science - Data Structures and Algorithms0202 electrical engineering electronic engineering information engineeringData Structures and Algorithms (cs.DS)Greedy algorithmFinite setAnalysis of algorithmsMathematicsSuperstring theoryShortest Common SuperstringComputer Science Applications1707 Computer Vision and Pattern RecognitionComputer Science ApplicationsReversalShortest Path Faster Algorithm010201 computation theory & mathematicsCompression ratioSignal Processing020201 artificial intelligence & image processingK shortest path routingInformation Systems
researchProduct

P2D: a self-supervised method for depth estimation from polarimetry

2021

Monocular depth estimation is a recurring subject in the field of computer vision. Its ability to describe scenes via a depth map while reducing the constraints related to the formulation of perspective geometry tends to favor its use. However, despite the constant improvement of algorithms, most methods exploit only colorimetric information. Consequently, robustness to events to which the modality is not sensitive to, like specularity or transparency, is neglected. In response to this phenomenon, we propose using polarimetry as an input for a self-supervised monodepth network. Therefore, we propose exploiting polarization cues to encourage accurate reconstruction of scenes. Furthermore, we…

FOS: Computer and information sciences0209 industrial biotechnologyMonocularComputer sciencebusiness.industryComputer Vision and Pattern Recognition (cs.CV)PolarimetryComputer Science - Computer Vision and Pattern RecognitionComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]02 engineering and technology010501 environmental sciences01 natural sciencesRegularization (mathematics)Term (time)020901 industrial engineering & automation[INFO.INFO-CV] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]SpecularityRobustness (computer science)Depth mapComputer visionArtificial intelligenceTransparency (data compression)business0105 earth and related environmental sciences
researchProduct

Novel Results on the Number of Runs of the Burrows-Wheeler-Transform

2021

The Burrows-Wheeler-Transform (BWT), a reversible string transformation, is one of the fundamental components of many current data structures in string processing. It is central in data compression, as well as in efficient query algorithms for sequence data, such as webpages, genomic and other biological sequences, or indeed any textual data. The BWT lends itself well to compression because its number of equal-letter-runs (usually referred to as $r$) is often considerably lower than that of the original string; in particular, it is well suited for strings with many repeated factors. In fact, much attention has been paid to the $r$ parameter as measure of repetitiveness, especially to evalua…

FOS: Computer and information sciencesBurrows–Wheeler transformSettore INF/01 - InformaticaCombinatorics on wordsFormal Languages and Automata Theory (cs.FL)Computer scienceString (computer science)Search engine indexingCompressed data structuresComputer Science - Formal Languages and Automata TheoryString indexingData structureMeasure (mathematics)Burrows-Wheeler-TransformRepetitivenessCombinatorics on wordsBurrows-Wheeler-Transform Compressed data structures String indexing Repetitiveness Combinatorics on wordsTransformation (function)Computer Science - Data Structures and AlgorithmsData Structures and Algorithms (cs.DS)AlgorithmData compression
researchProduct

Adaptive learning of compressible strings

2020

Suppose an oracle knows a string $S$ that is unknown to us and that we want to determine. The oracle can answer queries of the form "Is $s$ a substring of $S$?". In 1995, Skiena and Sundaram showed that, in the worst case, any algorithm needs to ask the oracle $\sigma n/4 -O(n)$ queries in order to be able to reconstruct the hidden string, where $\sigma$ is the size of the alphabet of $S$ and $n$ its length, and gave an algorithm that spends $(\sigma-1)n+O(\sigma \sqrt{n})$ queries to reconstruct $S$. The main contribution of our paper is to improve the above upper-bound in the context where the string is compressible. We first present a universal algorithm that, given a (computable) compre…

FOS: Computer and information sciencesCentroid decompositionGeneral Computer ScienceString compressionAdaptive learningKolmogorov complexityContext (language use)Data_CODINGANDINFORMATIONTHEORYString reconstructionTheoretical Computer ScienceCombinatoricsString reconstruction; String learning; Adaptive learning; Kolmogorov complexity; String compression; Lempel-Ziv; Centroid decomposition; Suffix treeSuffix treeIntegerComputer Science - Data Structures and AlgorithmsOrder (group theory)Data Structures and Algorithms (cs.DS)Adaptive learning; Centroid decomposition; Kolmogorov complexity; Lempel-Ziv; String compression; String learning; String reconstruction; Suffix treeTime complexityComputer Science::DatabasesMathematicsLempel-ZivSettore INF/01 - InformaticaLinear spaceString (computer science)SubstringBounded functionString learningTheoretical Computer Science
researchProduct

Uncommon Suffix Tries

2011

Common assumptions on the source producing the words inserted in a suffix trie with $n$ leaves lead to a $\log n$ height and saturation level. We provide an example of a suffix trie whose height increases faster than a power of $n$ and another one whose saturation level is negligible with respect to $\log n$. Both are built from VLMC (Variable Length Markov Chain) probabilistic sources; they are easily extended to families of sources having the same properties. The first example corresponds to a ''logarithmic infinite comb'' and enjoys a non uniform polynomial mixing. The second one corresponds to a ''factorial infinite comb'' for which mixing is uniform and exponential.

FOS: Computer and information sciencesCompressed suffix arrayPolynomialLogarithmGeneral MathematicsSuffix treevariable length Markov chain[INFO.INFO-DS]Computer Science [cs]/Data Structures and Algorithms [cs.DS]Generalized suffix treeprobabilistic source0102 computer and information sciences02 engineering and technologysuffix trie01 natural scienceslaw.inventionCombinatoricslawComputer Science - Data Structures and AlgorithmsTrieFOS: Mathematics0202 electrical engineering electronic engineering information engineeringData Structures and Algorithms (cs.DS)Mixing (physics)[ INFO.INFO-DS ] Computer Science [cs]/Data Structures and Algorithms [cs.DS]MathematicsDiscrete mathematicsApplied MathematicsProbability (math.PR)020206 networking & telecommunicationssuffix trie.Computer Graphics and Computer-Aided Design[MATH.MATH-PR]Mathematics [math]/Probability [math.PR]010201 computation theory & mathematicsmixing properties60J05 37E05Suffix[ MATH.MATH-PR ] Mathematics [math]/Probability [math.PR]Mathematics - ProbabilitySoftware
researchProduct

Perceptually Optimized Image Rendering

2017

We develop a framework for rendering photographic images by directly optimizing their perceptual similarity to the original visual scene. Specifically, over the set of all images that can be rendered on a given display, we minimize the normalized Laplacian pyramid distance (NLPD), a measure of perceptual dissimilarity that is derived from a simple model of the early stages of the human visual system. When rendering images acquired with a higher dynamic range than that of the display, we find that the optimization boosts the contrast of low-contrast features without introducing significant artifacts, yielding results of comparable visual quality to current state-of-the-art methods, but witho…

FOS: Computer and information sciencesComputer Science - Artificial IntelligenceImage qualityComputer scienceComputer Vision and Pattern Recognition (cs.CV)Computer Science - Computer Vision and Pattern RecognitionComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISIONImage processing02 engineering and technologyLuminanceRendering (computer graphics)Computer Science - GraphicsOptics0202 electrical engineering electronic engineering information engineeringComputer visionPower functionComputingMethodologies_COMPUTERGRAPHICSbusiness.industryDynamic range020207 software engineeringAtomic and Molecular Physics and OpticsGraphics (cs.GR)Electronic Optical and Magnetic MaterialsArtificial Intelligence (cs.AI)Human visual system model020201 artificial intelligence & image processingComputer Vision and Pattern RecognitionArtificial intelligencebusinessImage compression
researchProduct

Improving table compression with combinatorial optimization

2002

We study the problem of compressing massive tables within the partition-training paradigm introduced by Buchsbaum et al. [SODA'00], in which a table is partitioned by an off-line training procedure into disjoint intervals of columns, each of which is compressed separately by a standard, on-line compressor like gzip. We provide a new theory that unifies previous experimental observations on partitioning and heuristic observations on column permutation, all of which are used to improve compression rates. Based on the theory, we devise the first on-line training algorithms for table compression, which can be applied to individual files, not just continuously operating sources; and also a new, …

FOS: Computer and information sciencesComputer scienceHeuristic (computer science)E.4G.2.1Data_CODINGANDINFORMATIONTHEORYDisjoint setsTravelling salesman problemPermutationArtificial IntelligenceCompression (functional analysis)Computer Science - Data Structures and AlgorithmsH.1.8H.2.7Data Structures and Algorithms (cs.DS)E.4; F.1.3; F.2.2; G.2.1; H.1.1; H.1.8; H.2.7H.1.1Dynamic programmingHardware and ArchitectureControl and Systems EngineeringCombinatorial optimizationTable (database)F.1.3F.2.2AlgorithmSoftwareInformation SystemsJournal of the ACM
researchProduct

The rightmost equal-cost position problem.

2013

LZ77-based compression schemes compress the input text by replacing factors in the text with an encoded reference to a previous occurrence formed by the couple (length, offset). For a given factor, the smallest is the offset, the smallest is the resulting compression ratio. This is optimally achieved by using the rightmost occurrence of a factor in the previous text. Given a cost function, for instance the minimum number of bits used to represent an integer, we define the Rightmost Equal-Cost Position (REP) problem as the problem of finding one of the occurrences of a factor whose cost is equal to the cost of the rightmost one. We present the Multi-Layer Suffix Tree data structure that, for…

FOS: Computer and information sciencesOffset (computer science)Computer scienceSuffix treeComputer Science - Information Theorylaw.inventionCombinatoricslawLog-log plotComputer Science - Data Structures and AlgorithmsCompression schemetext compressiondictionary text compressionData Structures and Algorithms (cs.DS)LZ77 compressiondata compressionLossless compressionfull text indexSuffix Tree Data StructuresSettore INF/01 - InformaticaInformation Theory (cs.IT)Data structurePrefixCompression ratioCompression scheme; Constant time; Suffix Tree Data StructuresAlgorithmData compressionConstant time
researchProduct