|  | /* -*- linux-c -*- ------------------------------------------------------- * | 
|  | * | 
|  | *   Copyright 2001 H. Peter Anvin - All Rights Reserved | 
|  | * | 
|  | *   This program is free software; you can redistribute it and/or modify | 
|  | *   it under the terms of the GNU General Public License as published by | 
|  | *   the Free Software Foundation, Inc., 675 Mass Ave, Cambridge MA 02139, | 
|  | *   USA; either version 2 of the License, or (at your option) any later | 
|  | *   version; incorporated herein by reference. | 
|  | * | 
|  | * ----------------------------------------------------------------------- */ | 
|  |  | 
|  | /* | 
|  | * linux/fs/isofs/compress.c | 
|  | * | 
|  | * Transparent decompression of files on an iso9660 filesystem | 
|  | */ | 
|  |  | 
|  | #include <linux/module.h> | 
|  | #include <linux/init.h> | 
|  |  | 
|  | #include <linux/vmalloc.h> | 
|  | #include <linux/zlib.h> | 
|  |  | 
|  | #include "isofs.h" | 
|  | #include "zisofs.h" | 
|  |  | 
|  | /* This should probably be global. */ | 
|  | static char zisofs_sink_page[PAGE_CACHE_SIZE]; | 
|  |  | 
|  | /* | 
|  | * This contains the zlib memory allocation and the mutex for the | 
|  | * allocation; this avoids failures at block-decompression time. | 
|  | */ | 
|  | static void *zisofs_zlib_workspace; | 
|  | static DEFINE_MUTEX(zisofs_zlib_lock); | 
|  |  | 
|  | /* | 
|  | * When decompressing, we typically obtain more than one page | 
|  | * per reference.  We inject the additional pages into the page | 
|  | * cache as a form of readahead. | 
|  | */ | 
|  | static int zisofs_readpage(struct file *file, struct page *page) | 
|  | { | 
|  | struct inode *inode = file->f_path.dentry->d_inode; | 
|  | struct address_space *mapping = inode->i_mapping; | 
|  | unsigned int maxpage, xpage, fpage, blockindex; | 
|  | unsigned long offset; | 
|  | unsigned long blockptr, blockendptr, cstart, cend, csize; | 
|  | struct buffer_head *bh, *ptrbh[2]; | 
|  | unsigned long bufsize = ISOFS_BUFFER_SIZE(inode); | 
|  | unsigned int bufshift = ISOFS_BUFFER_BITS(inode); | 
|  | unsigned long bufmask  = bufsize - 1; | 
|  | int err = -EIO; | 
|  | int i; | 
|  | unsigned int header_size = ISOFS_I(inode)->i_format_parm[0]; | 
|  | unsigned int zisofs_block_shift = ISOFS_I(inode)->i_format_parm[1]; | 
|  | /* unsigned long zisofs_block_size = 1UL << zisofs_block_shift; */ | 
|  | unsigned int zisofs_block_page_shift = zisofs_block_shift-PAGE_CACHE_SHIFT; | 
|  | unsigned long zisofs_block_pages = 1UL << zisofs_block_page_shift; | 
|  | unsigned long zisofs_block_page_mask = zisofs_block_pages-1; | 
|  | struct page *pages[zisofs_block_pages]; | 
|  | unsigned long index = page->index; | 
|  | int indexblocks; | 
|  |  | 
|  | /* We have already been given one page, this is the one | 
|  | we must do. */ | 
|  | xpage = index & zisofs_block_page_mask; | 
|  | pages[xpage] = page; | 
|  |  | 
|  | /* The remaining pages need to be allocated and inserted */ | 
|  | offset = index & ~zisofs_block_page_mask; | 
|  | blockindex = offset >> zisofs_block_page_shift; | 
|  | maxpage = (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; | 
|  |  | 
|  | /* | 
|  | * If this page is wholly outside i_size we just return zero; | 
|  | * do_generic_file_read() will handle this for us | 
|  | */ | 
|  | if (page->index >= maxpage) { | 
|  | SetPageUptodate(page); | 
|  | unlock_page(page); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | maxpage = min(zisofs_block_pages, maxpage-offset); | 
|  |  | 
|  | for ( i = 0 ; i < maxpage ; i++, offset++ ) { | 
|  | if ( i != xpage ) { | 
|  | pages[i] = grab_cache_page_nowait(mapping, offset); | 
|  | } | 
|  | page = pages[i]; | 
|  | if ( page ) { | 
|  | ClearPageError(page); | 
|  | kmap(page); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* This is the last page filled, plus one; used in case of abort. */ | 
|  | fpage = 0; | 
|  |  | 
|  | /* Find the pointer to this specific chunk */ | 
|  | /* Note: we're not using isonum_731() here because the data is known aligned */ | 
|  | /* Note: header_size is in 32-bit words (4 bytes) */ | 
|  | blockptr = (header_size + blockindex) << 2; | 
|  | blockendptr = blockptr + 4; | 
|  |  | 
|  | indexblocks = ((blockptr^blockendptr) >> bufshift) ? 2 : 1; | 
|  | ptrbh[0] = ptrbh[1] = NULL; | 
|  |  | 
|  | if ( isofs_get_blocks(inode, blockptr >> bufshift, ptrbh, indexblocks) != indexblocks ) { | 
|  | if ( ptrbh[0] ) brelse(ptrbh[0]); | 
|  | printk(KERN_DEBUG "zisofs: Null buffer on reading block table, inode = %lu, block = %lu\n", | 
|  | inode->i_ino, blockptr >> bufshift); | 
|  | goto eio; | 
|  | } | 
|  | ll_rw_block(READ, indexblocks, ptrbh); | 
|  |  | 
|  | bh = ptrbh[0]; | 
|  | if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) { | 
|  | printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n", | 
|  | inode->i_ino, blockptr >> bufshift); | 
|  | if ( ptrbh[1] ) | 
|  | brelse(ptrbh[1]); | 
|  | goto eio; | 
|  | } | 
|  | cstart = le32_to_cpu(*(__le32 *)(bh->b_data + (blockptr & bufmask))); | 
|  |  | 
|  | if ( indexblocks == 2 ) { | 
|  | /* We just crossed a block boundary.  Switch to the next block */ | 
|  | brelse(bh); | 
|  | bh = ptrbh[1]; | 
|  | if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) { | 
|  | printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n", | 
|  | inode->i_ino, blockendptr >> bufshift); | 
|  | goto eio; | 
|  | } | 
|  | } | 
|  | cend = le32_to_cpu(*(__le32 *)(bh->b_data + (blockendptr & bufmask))); | 
|  | brelse(bh); | 
|  |  | 
|  | if (cstart > cend) | 
|  | goto eio; | 
|  |  | 
|  | csize = cend-cstart; | 
|  |  | 
|  | if (csize > deflateBound(1UL << zisofs_block_shift)) | 
|  | goto eio; | 
|  |  | 
|  | /* Now page[] contains an array of pages, any of which can be NULL, | 
|  | and the locks on which we hold.  We should now read the data and | 
|  | release the pages.  If the pages are NULL the decompressed data | 
|  | for that particular page should be discarded. */ | 
|  |  | 
|  | if ( csize == 0 ) { | 
|  | /* This data block is empty. */ | 
|  |  | 
|  | for ( fpage = 0 ; fpage < maxpage ; fpage++ ) { | 
|  | if ( (page = pages[fpage]) != NULL ) { | 
|  | memset(page_address(page), 0, PAGE_CACHE_SIZE); | 
|  |  | 
|  | flush_dcache_page(page); | 
|  | SetPageUptodate(page); | 
|  | kunmap(page); | 
|  | unlock_page(page); | 
|  | if ( fpage == xpage ) | 
|  | err = 0; /* The critical page */ | 
|  | else | 
|  | page_cache_release(page); | 
|  | } | 
|  | } | 
|  | } else { | 
|  | /* This data block is compressed. */ | 
|  | z_stream stream; | 
|  | int bail = 0, left_out = -1; | 
|  | int zerr; | 
|  | int needblocks = (csize + (cstart & bufmask) + bufmask) >> bufshift; | 
|  | int haveblocks; | 
|  | struct buffer_head *bhs[needblocks+1]; | 
|  | struct buffer_head **bhptr; | 
|  |  | 
|  | /* Because zlib is not thread-safe, do all the I/O at the top. */ | 
|  |  | 
|  | blockptr = cstart >> bufshift; | 
|  | memset(bhs, 0, (needblocks+1)*sizeof(struct buffer_head *)); | 
|  | haveblocks = isofs_get_blocks(inode, blockptr, bhs, needblocks); | 
|  | ll_rw_block(READ, haveblocks, bhs); | 
|  |  | 
|  | bhptr = &bhs[0]; | 
|  | bh = *bhptr++; | 
|  |  | 
|  | /* First block is special since it may be fractional. | 
|  | We also wait for it before grabbing the zlib | 
|  | mutex; odds are that the subsequent blocks are | 
|  | going to come in in short order so we don't hold | 
|  | the zlib mutex longer than necessary. */ | 
|  |  | 
|  | if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) { | 
|  | printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n", | 
|  | fpage, xpage, csize); | 
|  | goto b_eio; | 
|  | } | 
|  | stream.next_in  = bh->b_data + (cstart & bufmask); | 
|  | stream.avail_in = min(bufsize-(cstart & bufmask), csize); | 
|  | csize -= stream.avail_in; | 
|  |  | 
|  | stream.workspace = zisofs_zlib_workspace; | 
|  | mutex_lock(&zisofs_zlib_lock); | 
|  |  | 
|  | zerr = zlib_inflateInit(&stream); | 
|  | if ( zerr != Z_OK ) { | 
|  | if ( err && zerr == Z_MEM_ERROR ) | 
|  | err = -ENOMEM; | 
|  | printk(KERN_DEBUG "zisofs: zisofs_inflateInit returned %d\n", | 
|  | zerr); | 
|  | goto z_eio; | 
|  | } | 
|  |  | 
|  | while ( !bail && fpage < maxpage ) { | 
|  | page = pages[fpage]; | 
|  | if ( page ) | 
|  | stream.next_out = page_address(page); | 
|  | else | 
|  | stream.next_out = (void *)&zisofs_sink_page; | 
|  | stream.avail_out = PAGE_CACHE_SIZE; | 
|  |  | 
|  | while ( stream.avail_out ) { | 
|  | int ao, ai; | 
|  | if ( stream.avail_in == 0 && left_out ) { | 
|  | if ( !csize ) { | 
|  | printk(KERN_WARNING "zisofs: ZF read beyond end of input\n"); | 
|  | bail = 1; | 
|  | break; | 
|  | } else { | 
|  | bh = *bhptr++; | 
|  | if ( !bh || | 
|  | (wait_on_buffer(bh), !buffer_uptodate(bh)) ) { | 
|  | /* Reached an EIO */ | 
|  | printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n", | 
|  | fpage, xpage, csize); | 
|  |  | 
|  | bail = 1; | 
|  | break; | 
|  | } | 
|  | stream.next_in = bh->b_data; | 
|  | stream.avail_in = min(csize,bufsize); | 
|  | csize -= stream.avail_in; | 
|  | } | 
|  | } | 
|  | ao = stream.avail_out;  ai = stream.avail_in; | 
|  | zerr = zlib_inflate(&stream, Z_SYNC_FLUSH); | 
|  | left_out = stream.avail_out; | 
|  | if ( zerr == Z_BUF_ERROR && stream.avail_in == 0 ) | 
|  | continue; | 
|  | if ( zerr != Z_OK ) { | 
|  | /* EOF, error, or trying to read beyond end of input */ | 
|  | if ( err && zerr == Z_MEM_ERROR ) | 
|  | err = -ENOMEM; | 
|  | if ( zerr != Z_STREAM_END ) | 
|  | printk(KERN_DEBUG "zisofs: zisofs_inflate returned %d, inode = %lu, index = %lu, fpage = %d, xpage = %d, avail_in = %d, avail_out = %d, ai = %d, ao = %d\n", | 
|  | zerr, inode->i_ino, index, | 
|  | fpage, xpage, | 
|  | stream.avail_in, stream.avail_out, | 
|  | ai, ao); | 
|  | bail = 1; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | if ( stream.avail_out && zerr == Z_STREAM_END ) { | 
|  | /* Fractional page written before EOF.  This may | 
|  | be the last page in the file. */ | 
|  | memset(stream.next_out, 0, stream.avail_out); | 
|  | stream.avail_out = 0; | 
|  | } | 
|  |  | 
|  | if ( !stream.avail_out ) { | 
|  | /* This page completed */ | 
|  | if ( page ) { | 
|  | flush_dcache_page(page); | 
|  | SetPageUptodate(page); | 
|  | kunmap(page); | 
|  | unlock_page(page); | 
|  | if ( fpage == xpage ) | 
|  | err = 0; /* The critical page */ | 
|  | else | 
|  | page_cache_release(page); | 
|  | } | 
|  | fpage++; | 
|  | } | 
|  | } | 
|  | zlib_inflateEnd(&stream); | 
|  |  | 
|  | z_eio: | 
|  | mutex_unlock(&zisofs_zlib_lock); | 
|  |  | 
|  | b_eio: | 
|  | for ( i = 0 ; i < haveblocks ; i++ ) { | 
|  | if ( bhs[i] ) | 
|  | brelse(bhs[i]); | 
|  | } | 
|  | } | 
|  |  | 
|  | eio: | 
|  |  | 
|  | /* Release any residual pages, do not SetPageUptodate */ | 
|  | while ( fpage < maxpage ) { | 
|  | page = pages[fpage]; | 
|  | if ( page ) { | 
|  | flush_dcache_page(page); | 
|  | if ( fpage == xpage ) | 
|  | SetPageError(page); | 
|  | kunmap(page); | 
|  | unlock_page(page); | 
|  | if ( fpage != xpage ) | 
|  | page_cache_release(page); | 
|  | } | 
|  | fpage++; | 
|  | } | 
|  |  | 
|  | /* At this point, err contains 0 or -EIO depending on the "critical" page */ | 
|  | return err; | 
|  | } | 
|  |  | 
|  | const struct address_space_operations zisofs_aops = { | 
|  | .readpage = zisofs_readpage, | 
|  | /* No sync_page operation supported? */ | 
|  | /* No bmap operation supported */ | 
|  | }; | 
|  |  | 
|  | int __init zisofs_init(void) | 
|  | { | 
|  | zisofs_zlib_workspace = vmalloc(zlib_inflate_workspacesize()); | 
|  | if ( !zisofs_zlib_workspace ) | 
|  | return -ENOMEM; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void zisofs_cleanup(void) | 
|  | { | 
|  | vfree(zisofs_zlib_workspace); | 
|  | } |