2,748 research outputs found

    Work design improvement at Miroad Rubber Industries Sdn. Bhd.

    Get PDF
    Erul Food Industries known as Salaiport Industry is a family-owned company and was established on July 2017. Salaiport Industry apparently moved to a new place at Pedas, Negeri Sembilan. Previously, Salaiport Industry operated in-house located at Pagoh, Johor. This small company major business is producing frozen smoked beef, smoked quail, smoke catfish and smoked duck. The main frozen product is smoked beef. The frozen smoked meat produced by Salaiport Industry is depending on customer demands. Usually the company produce 40 kg to 60 kg a day and operated between for four days until five days. Therefore, the company produce approximately around 80 kg to 120 kg per week. The company usually take 2 days for 1 complete cycle for the production as the first day the company will only receive the meat from the supplier and freeze the meat for use of tomorrow

    Image Characterization and Classification by Physical Complexity

    Full text link
    We present a method for estimating the complexity of an image based on Bennett's concept of logical depth. Bennett identified logical depth as the appropriate measure of organized complexity, and hence as being better suited to the evaluation of the complexity of objects in the physical world. Its use results in a different, and in some sense a finer characterization than is obtained through the application of the concept of Kolmogorov complexity alone. We use this measure to classify images by their information content. The method provides a means for classifying and evaluating the complexity of objects by way of their visual representations. To the authors' knowledge, the method and application inspired by the concept of logical depth presented herein are being proposed and implemented for the first time.Comment: 30 pages, 21 figure

    Multi-rate, real time image compression for images dominated by point sources

    Get PDF
    An image compression system recently developed for compression of digital images dominated by point sources is presented. Encoding consists of minimum-mean removal, vector quantization, adaptive threshold truncation, and modified Huffman encoding. Simulations are presented showing that the peaks corresponding to point sources can be transmitted losslessly for low signal-to-noise ratios (SNR) and high point source densities while maintaining a reduced output bit rate. Encoding and decoding hardware has been built and tested which processes 552,960 12-bit pixels per second at compression rates of 10:1 and 4:1. Simulation results are presented for the 10:1 case only

    Compression of spectral meteorological imagery

    Get PDF
    Data compression is essential to current low-earth-orbit spectral sensors with global coverage, e.g., meteorological sensors. Such sensors routinely produce in excess of 30 Gb of data per orbit (over 4 Mb/s for about 110 min) while typically limited to less than 10 Gb of downlink capacity per orbit (15 minutes at 10 Mb/s). Astro-Space Division develops spaceborne compression systems for compression ratios from as little as three to as much as twenty-to-one for high-fidelity reconstructions. Current hardware production and development at Astro-Space Division focuses on discrete cosine transform (DCT) systems implemented with the GE PFFT chip, a 32x32 2D-DCT engine. Spectral relations in the data are exploited through block mean extraction followed by orthonormal transformation. The transformation produces blocks with spatial correlation that are suitable for further compression with any block-oriented spatial compression system, e.g., Astro-Space Division's Laplacian modeler and analytic encoder of DCT coefficients

    A very high speed lossless compression/decompression chip set

    Get PDF
    A chip is described that will perform lossless compression and decompression using the Rice Algorithm. The chip set is designed to compress and decompress source data in real time for many applications. The encoder is designed to code at 20 M samples/second at MIL specifications. That corresponds to 280 Mbits/second at maximum quantization or approximately 500 Mbits/second under nominal conditions. The decoder is designed to decode at 10 M samples/second at industrial specifications. A wide range of quantization levels is allowed (4...14 bits) and both nearest neighbor prediction and external prediction are supported. When the pre and post processors are bypassed, the chip set performs high speed entropy coding and decoding. This frees the chip set from being tied to one modeling technique or specific application. Both the encoder and decoder are being fabricated in a 1.0 micron CMOS process that has been tested to survive 1 megarad of total radiation dosage. The CMOS chips are small, only 5 mm on a side, and both are estimated to consume less than 1/4 of a Watt of power while operating at maximum frequency

    Compression and Classification Methods for Galaxy Spectra in Large Redshift Surveys

    Get PDF
    Methods for compression and classification of galaxy spectra, which are useful for large galaxy redshift surveys (such as the SDSS, 2dF, 6dF and VIRMOS), are reviewed. In particular, we describe and contrast three methods: (i) Principal Component Analysis, (ii) Information Bottleneck, and (iii) Fisher Matrix. We show applications to 2dF galaxy spectra and to mock semi-analytic spectra, and we discuss how these methods can be used to study physical processes of galaxy formation, clustering and galaxy biasing in the new large redshift surveys.Comment: Review talk, proceedings of MPA/MPE/ESO Conference "Mining the Sky", 2000, Garching, Germany; 20 pages, 5 figure

    Efficient Data Compression with Error Bound Guarantee in Wireless Sensor Networks

    Get PDF
    We present a data compression and dimensionality reduction scheme for data fusion and aggregation applications to prevent data congestion and reduce energy consumption at network connecting points such as cluster heads and gateways. Our in-network approach can be easily tuned to analyze the data temporal or spatial correlation using an unsupervised neural network scheme, namely the autoencoders. In particular, our algorithm extracts intrinsic data features from previously collected historical samples to transform the raw data into a low dimensional representation. Moreover, the proposed framework provides an error bound guarantee mechanism. We evaluate the proposed solution using real-world data sets and compare it with traditional methods for temporal and spatial data compression. The experimental validation reveals that our approach outperforms several existing wireless sensor network's data compression methods in terms of compression efficiency and signal reconstruction.Comment: ACM MSWiM 201
    corecore