Skip to content

a streaming interface for archive generation

License

Notifications You must be signed in to change notification settings

solcates/node-archiver

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Archiver v0.15.0 Build Status Build status

a streaming interface for archive generation

NPM

Install

npm install archiver --save

Usage

var archiver = require('archiver');
var archive = archiver.create('zip', {}); // or archiver('zip', {});

API

Transform

Inherits Transform Stream methods.

create(format, options)

Creates an Archiver instance based on the format (zip, tar, etc) passed. Parameters can be passed directly to Archiver constructor for convenience.

abort()

Aborts the archiving process, taking a best-effort approach, by:

  • removing any pending queue tasks
  • allowing any active queue workers to finish
  • detaching internal module pipes
  • ending both sides of the Transform stream

It will NOT drain any remaining sources.

append(input, data)

Appends an input source (text string, buffer, or stream) to the instance. When the instance has received, processed, and emitted the input, the entry event is fired.

Replaced #addFile in v0.5.

archive.append('string', { name:'string.txt' });
archive.append(new Buffer('string'), { name:'buffer.txt' });
archive.append(fs.createReadStream('mydir/file.txt'), { name:'stream.txt' });
archive.append(null, { name:'dir/' });

bulk(mappings)

Appends multiple entries from passed array of src-dest mappings. A lazystream wrapper is used to prevent issues with open file limits.

Globbing patterns are supported through use of the bundled file-utils module.

The data property can be set (per src-dest mapping) to define data for matched entries.

archive.bulk([
  { src: ['mydir/**'], data: { date: new Date() } },
  { src: ['mydir/**'], data: function(data) {
    data.date = new Date();
    return data;
  }},
  { expand: true, cwd: 'mydir', src: ['**'], dest: 'newdir' }
]);

As of v0.15, the data property can also be a function that receives data for each matched entry and is expected to return it after making any desired adjustments.

For more detail on this feature, please see BULK.md.

directory(dirpath[, destpath, data])

Appends a directory and its files, recusively, given its dirpath. This is meant to be a simplier approach to something previously only possible with bulk. The use of destpath allows one to define a custom destination path within the resulting archive and data allows for setting data on each entry appended.

// mydir/ -> archive.ext/mydir/
archive.directory('mydir');

// mydir/ -> archive.ext/abc/
archive.directory('mydir', 'abc');

// mydir/ -> archive.ext/
archive.directory('mydir', false, { date: new Date() });
archive.directory('mydir', false, function(data) {
  data.date = new Date();
  return data;
});

As of v0.15, the data property can also be a function that receives data for each entry and is expected to return it after making any desired adjustments.

file(filepath, data)

Appends a file given its filepath using a lazystream wrapper to prevent issues with open file limits. When the instance has received, processed, and emitted the file, the entry event is fired.

archive.file('mydir/file.txt', { name:'file.txt' });

finalize()

Finalizes the instance and prevents further appending to the archive structure (queue will continue til drained). The end, close or finish events on the destination stream may fire right after calling this method so you should set listeners beforehand to properly detect stream completion.

You must call this method to get a valid archive and end the instance stream.

pointer()

Returns the current byte length emitted by archiver. Use this in your end callback to log generated size.

use(plugin)

Add a plugin to the middleware stack. Currently this is designed for passing the module to use (replaces registerFormat/setFormat/setModule)

Events

Inherits Transform Stream events.

entry

Fired when the entry's input has been processed and appended to the archive. Passes entry data as first argument.

Zip

Options

comment string

Sets the zip comment.

statConcurrency number

Sets the number of workers used to process the internal fs stat queue. Defaults to 4.

store boolean

If true, all entries will be archived without compression. Defaults to false.

zlib object

Passed to node's zlib module to control compression. Options may vary by node version.

Entry Data

name string required

Sets the entry name including internal path.

date string|Date

Sets the entry date. This can be any valid date string or instance. Defaults to current time in locale.

When using the bulk or file methods, fs stat data is used as the default value.

store boolean

If true, this entry will be archived without compression. Defaults to global store option.

comment string

Sets the entry comment.

mode number

Sets the entry permissions. Defaults to octal 0755 (directory) or 0644 (file).

When using the bulk or file methods, fs stat data is used as the default value.

stats fs.Stats

Sets the fs stat data for this entry. This allows for reduction of fs stat calls when stat data is already known.

Tar

Options

gzip boolean

Compresses the tar archive using gzip, default is false.

gzipOptions object

Passed to node's zlib module to control compression. Options may vary by node version.

statConcurrency number

Sets the number of workers used to process the internal fs stat queue. Defaults to 4.

Entry Data

name string required

Sets the entry name including internal path.

date string|Date

Sets the entry date. This can be any valid date string or instance. Defaults to current time in locale.

When using the bulk or file methods, fs stat data is used as the default value.

mode number

Sets the entry permissions. Defaults to octal 0755 (directory) or 0644 (file).

When using the bulk or file methods, fs stat data is used as the default value.

stats fs.Stats

Sets the fs stat data for this entry. This allows for reduction of fs stat calls when stat data is already known.

Custom Formats

Archiver ships with out of the box support for TAR and ZIP archives.

You can register additional formats with registerFormat.

Formats will be changing in the next few releases to implement a middleware approach.

Libraries

Archiver makes use of several libraries/modules to avoid duplication of efforts.

Things of Interest

About

a streaming interface for archive generation

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • JavaScript 100.0%