• Title/Summary/Keyword: Chunk

Search Result 109, Processing Time 0.026 seconds

Enhancement of SCTP Throughput using Chunk Checksum

  • Lin Cui;Koh Seok J.;Hong Yong-Geun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2006.05a
    • /
    • pp.1147-1150
    • /
    • 2006
  • Stream Control Transmission Protocol (SCTP) uses the 32-bit checksum in the common header, by which a corrupted SCTP packet will be regarded as a lost packet and then discarded. This may result in degradation of SCTP's throughput performance over wireless networks. This paper proposes a new chunk checksum scheme for SCTP, in which each data chunk contains its own checksum field and SACK chunk carry corresponding Transmission Sequence Number (TSN) with timestamp for every corruption event. The proposed chunk checksum scheme is introduced with the following three purposes: 1) to distinguish the chunk corruptions from the chunk losses; 2) to avoid the unnecessary halving of the congestion window (cwnd) in the case of chunk corruption; 3) to avoid the unwanted timeouts which can be induced in conventional SCTP in the case that the retransmitted data chunks are corrupted again in wireless networks. Simulation results show that the proposed chunk checksum scheme could improve the SCTP throughput in the wireless environments with a high bit error rate.

  • PDF

A Parallel Loop Scheduling Algorithm on Multiprocessor System Environments (다중프로세서 시스템 환경에서 병렬 루프 스케쥴링 알고리즘)

  • 이영규;박두순
    • Journal of Korea Multimedia Society
    • /
    • v.3 no.3
    • /
    • pp.309-319
    • /
    • 2000
  • The purpose of a parallel scheduling under a multiprocessor environment is to carry out the scheduling with the minimum synchronization overhead, and to perform load balance for a parallel application program. The processors calculate the chunk of iteration and are allocated to carry out the parallel iteration. At this time, it frequently accesses mutually exclusive global memory so that there are a lot of scheduling overhead and bottleneck imposed. And also, when the distribution of the parallel iteration in the allocated chunk to the processor is different, the different execution time of each chunk causes the load imbalance and badly affects the capability of the all scheduling. In the paper. we investigate the problems on the conventional algorithms in order to achieve the minimum scheduling overhead and load balance. we then present a new parallel loop scheduling algorithm, considering the locality of the data and processor affinity.

  • PDF

Efficient Content Delivery Method in Wireless Content-Centric Network (무선 Content-Centric Network에서 효과적인 콘텐츠 전달 방식)

  • Park, Chan-Min;Kim, Byung-Seo
    • Journal of Internet Computing and Services
    • /
    • v.18 no.2
    • /
    • pp.13-20
    • /
    • 2017
  • Recently, researches to adopt Content-centric network (CCN), which is one of the promising technologies for replacing TCP/IP-based networks, to wireless networks has actively performed. However, because of erroneous and unreliable channel characteristics, there are many problems to be resolved to adopt CCN to wireless networks. This paper proposes a method to reduce content download time because nodes possess only parts of content chunks. The proposed method enables a node having parts of content chunks to request the rest of parts of content chunks to a provider before a Consumer requests the content. As a consequence, the content download time is reduced.

Dynamic Prime Chunking Algorithm for Data Deduplication in Cloud Storage

  • Ellappan, Manogar;Abirami, S
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.4
    • /
    • pp.1342-1359
    • /
    • 2021
  • The data deduplication technique identifies the duplicates and minimizes the redundant storage data in the backup server. The chunk level deduplication plays a significant role in detecting the appropriate chunk boundaries, which solves the challenges such as minimum throughput and maximum chunk size variance in the data stream. To provide the solution, we propose a new chunking algorithm called Dynamic Prime Chunking (DPC). The main goal of DPC is to dynamically change the window size within the prime value based on the minimum and maximum chunk size. According to the result, DPC provides high throughput and avoid significant chunk variance in the deduplication system. The implementation and experimental evaluation have been performed on the multimedia and operating system datasets. DPC has been compared with existing algorithms such as Rabin, TTTD, MAXP, and AE. Chunk Count, Chunking time, throughput, processing time, Bytes Saved per Second (BSPS) and Deduplication Elimination Ratio (DER) are the performance metrics analyzed in our work. Based on the analysis of the results, it is found that throughput and BSPS have improved. Firstly, DPC quantitatively improves throughput performance by more than 21% than AE. Secondly, BSPS increases a maximum of 11% than the existing AE algorithm. Due to the above reason, our algorithm minimizes the total processing time and achieves higher deduplication efficiency compared with the existing Content Defined Chunking (CDC) algorithms.

Transaction Update method based on Fuzzy Chunk for QoS Performance Improvement of Mobile Streaming Service (모바일 스트리밍 서비스의 QoS 개선을 위한 퍼지 청크 기반의 트랜잭션 갱신방법)

  • Jeong, Taeg-Won;Lee, Chong-Deuk
    • Journal of Digital Contents Society
    • /
    • v.9 no.4
    • /
    • pp.543-550
    • /
    • 2008
  • There exist many QoS problems such as delay and jitter for mobile streaming due to limited bandwidth. This paper proposed a transaction update method based on Fuzzy chunk to solve the problems. The proposed method updated a transaction, which is classified as a read or an update, using Fuzzy filtering. A chunk block is rearranged according to the fuzziness which is classified as better relevance for fuzziness greater than 0.8, relevance for fuzziness between 0.5 and 0.7, and less irrelevance for fuzziness less than 0.5. According to the simulation result for the rearranged chunk block, the proposed method has better performance than those of RBM(Randon Based method), DBM(Distance Based Method), and FBM(Frequency Based Method).

  • PDF

Speed Up of OLAP Query response time By Chunk-based Caching Scheme In Shared nothing Structure (Shared nothing 구조하에서 Chunk-based Caching 전략에 의한 OLAP Query response time의 속도향상)

  • 엄준식;정병수
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10a
    • /
    • pp.135-137
    • /
    • 1999
  • 현재 의사결정 시스템을 위한 데이터 웨어하우스는 데이터베이스 분야에서 비약적인 발전을 해오고 있다. 이 분야에서 중요한 사항은 사용자의 질의에 대한 응답시간이다. 사용자로부터 질의가 요구되면 과거의 많은 데이터를 분석하여 적절한 시간 내에 분석결과를 되돌려 줄 수 있어야 하는데 데이터 웨어하우스의 특성상 대용량의 데이터를 저장하고 분석 시에 많은 데이터를 검색해야 하므로 질의 응답시간에 많은 시간이 소요된다. 이 논문에서는 Chunk based caching 기법에 의해서 새로이 요구되는 질의에 대해 이미 메모리에 캐시 되어진 내용을 이용하는 방식을 통해 디스크의 I/O 횟수를 줄임으로서 질의 응답시간을 단축시키는 기법을 소개한다. 또한 chunk miss에 대한 처리를 병렬로 수행함으로써 메모리에 캐시되지 않은 내용을 디스크로부터 로드하는 시간을 단축시키는 방법도 아울러 소개한다.

  • PDF

Effectively Using Memory throughout Adjustment for Slab Chunk Size of Memcached (memcached의 slab chunk 크기 조정을 통한 효율적인 메모리 사용)

  • Son, Ju-Hyoung;Lee, MinJae;Kim, Tae Il;Kang, SooYong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.11a
    • /
    • pp.1399-1401
    • /
    • 2012
  • memcached는 기존 RDBMS의 한계를 극복하기 위해 나온 소프트웨어 중 하나이다. memcached는 여러 장점들이 있어 많은 곳에서 활용되지만 주의할 점이 있어 사용 시 몇 가지 요소들을 적절히 설정하여 사용해야 한다. memcached는 기존메모리 할당방식 대신 slab allocater를 통해 메모리를 관리하여 입력되는 데이터의 크기에 맞춰 slab class의 chunk에 저장된다. 입력되는 데이터의 크기에 따라 저장된다는 memcached의 특성 때문에 slab chunk 크기를 조정하는 요소를 상황에 맞게 정해준다면 보다 효율적인 메모리 사용을 할 수 있다는 것을 실험을 통해 검증해보았고, 이 실험을 통해 나온 결과는 memcached를 사용하는 많은 분야에서 응용될 수 있을 것이다.

Research on Keyword-Overlap Similarity Algorithm Optimization in Short English Text Based on Lexical Chunk Theory

  • Na Li;Cheng Li;Honglie Zhang
    • Journal of Information Processing Systems
    • /
    • v.19 no.5
    • /
    • pp.631-640
    • /
    • 2023
  • Short-text similarity calculation is one of the hot issues in natural language processing research. The conventional keyword-overlap similarity algorithms merely consider the lexical item information and neglect the effect of the word order. And some of its optimized algorithms combine the word order, but the weights are hard to be determined. In the paper, viewing the keyword-overlap similarity algorithm, the short English text similarity algorithm based on lexical chunk theory (LC-SETSA) is proposed, which introduces the lexical chunk theory existing in cognitive psychology category into the short English text similarity calculation for the first time. The lexical chunks are applied to segment short English texts, and the segmentation results demonstrate the semantic connotation and the fixed word order of the lexical chunks, and then the overlap similarity of the lexical chunks is calculated accordingly. Finally, the comparative experiments are carried out, and the experimental results prove that the proposed algorithm of the paper is feasible, stable, and effective to a large extent.

Effect of Packaging Method on the Storage Stability of Hair Tail Products (포장방법이 칼치제품의 저장성에 미치는 영향)

  • Jo, Kil-Suk;Kim, Hyun-Ku;Kang, Tong-Sam;Shin, Dong-Hwa
    • Korean Journal of Food Science and Technology
    • /
    • v.20 no.1
    • /
    • pp.45-51
    • /
    • 1988
  • To improve the individual packaging method and extend the shelf life of hair tail(Trichiurus japonicus), salted an unsalted hair tail chunk (cut in 8-10cm) were packaged in laminated plastic film bag(Nylon/PE: $20{\mu}m,\;12{\times}15cm$) filled with with free-$O_2$ absorber, in vacuum, and stored at 0 and/or $5^{\circ}C$. The other samples were packaged in plastic foam trays, overwraped with oxygen permeable film(control), and stored at same temperature. Volatile basic nitrogen (VBN), trimethylamine (TMA) and viable cell counts (VCC) were progressed with increase of storage time, but thiobarbituric acid (TBA) values decreased gradually after reaching at a maximum peak in 5-15 days. Judging from 4 chemical components, VBN was the most available component in quality judgement of hair tail chunk and its upper limiting content was 29 mg%. Regression equation for shelf life prediction of hair tail chunk with sensory evalution and VBN component was determined.

  • PDF

A Bitmap Index for Chunk-Based MOLAP Cubes (청크 기반 MOLAP 큐브를 위한 비트맵 인덱스)

  • Lim, Yoon-Sun;Kim, Myung
    • Journal of KIISE:Databases
    • /
    • v.30 no.3
    • /
    • pp.225-236
    • /
    • 2003
  • MOLAP systems store data in a multidimensional away called a 'cube' and access them using way indexes. When a cube is placed into disk, it can be Partitioned into a set of chunks of the same side length. Such a cube storage scheme is called the chunk-based MOLAP cube storage scheme. It gives data clustering effect so that all the dimensions are guaranteed to get a fair chance in terms of the query processing speed. In order to achieve high space utilization, sparse chunks are further compressed. Due to data compression, the relative position of chunks cannot be obtained in constant time without using indexes. In this paper, we propose a bitmap index for chunk-based MOLAP cubes. The index can be constructed along with the corresponding cube generation. The relative position of chunks is retained in the index so that chunk retrieval can be done in constant time. We placed in an index block as many chunks as possible so that the number of index searches is minimized for OLAP operations such as range queries. We showed the proposed index is efficient by comparing it with multidimensional indexes such as UB-tree and grid file in terms of time and space.