You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					510 lines
				
				14 KiB
			
		
		
			
		
	
	
					510 lines
				
				14 KiB
			| 
											3 years ago
										 | 'use strict' | ||
|  | 
 | ||
|  | // this[BUFFER] is the remainder of a chunk if we're waiting for
 | ||
|  | // the full 512 bytes of a header to come in.  We will Buffer.concat()
 | ||
|  | // it to the next write(), which is a mem copy, but a small one.
 | ||
|  | //
 | ||
|  | // this[QUEUE] is a Yallist of entries that haven't been emitted
 | ||
|  | // yet this can only get filled up if the user keeps write()ing after
 | ||
|  | // a write() returns false, or does a write() with more than one entry
 | ||
|  | //
 | ||
|  | // We don't buffer chunks, we always parse them and either create an
 | ||
|  | // entry, or push it into the active entry.  The ReadEntry class knows
 | ||
|  | // to throw data away if .ignore=true
 | ||
|  | //
 | ||
|  | // Shift entry off the buffer when it emits 'end', and emit 'entry' for
 | ||
|  | // the next one in the list.
 | ||
|  | //
 | ||
|  | // At any time, we're pushing body chunks into the entry at WRITEENTRY,
 | ||
|  | // and waiting for 'end' on the entry at READENTRY
 | ||
|  | //
 | ||
|  | // ignored entries get .resume() called on them straight away
 | ||
|  | 
 | ||
|  | const warner = require('./warn-mixin.js') | ||
|  | const Header = require('./header.js') | ||
|  | const EE = require('events') | ||
|  | const Yallist = require('yallist') | ||
|  | const maxMetaEntrySize = 1024 * 1024 | ||
|  | const Entry = require('./read-entry.js') | ||
|  | const Pax = require('./pax.js') | ||
|  | const zlib = require('minizlib') | ||
|  | const { nextTick } = require('process') | ||
|  | 
 | ||
|  | const gzipHeader = Buffer.from([0x1f, 0x8b]) | ||
|  | const STATE = Symbol('state') | ||
|  | const WRITEENTRY = Symbol('writeEntry') | ||
|  | const READENTRY = Symbol('readEntry') | ||
|  | const NEXTENTRY = Symbol('nextEntry') | ||
|  | const PROCESSENTRY = Symbol('processEntry') | ||
|  | const EX = Symbol('extendedHeader') | ||
|  | const GEX = Symbol('globalExtendedHeader') | ||
|  | const META = Symbol('meta') | ||
|  | const EMITMETA = Symbol('emitMeta') | ||
|  | const BUFFER = Symbol('buffer') | ||
|  | const QUEUE = Symbol('queue') | ||
|  | const ENDED = Symbol('ended') | ||
|  | const EMITTEDEND = Symbol('emittedEnd') | ||
|  | const EMIT = Symbol('emit') | ||
|  | const UNZIP = Symbol('unzip') | ||
|  | const CONSUMECHUNK = Symbol('consumeChunk') | ||
|  | const CONSUMECHUNKSUB = Symbol('consumeChunkSub') | ||
|  | const CONSUMEBODY = Symbol('consumeBody') | ||
|  | const CONSUMEMETA = Symbol('consumeMeta') | ||
|  | const CONSUMEHEADER = Symbol('consumeHeader') | ||
|  | const CONSUMING = Symbol('consuming') | ||
|  | const BUFFERCONCAT = Symbol('bufferConcat') | ||
|  | const MAYBEEND = Symbol('maybeEnd') | ||
|  | const WRITING = Symbol('writing') | ||
|  | const ABORTED = Symbol('aborted') | ||
|  | const DONE = Symbol('onDone') | ||
|  | const SAW_VALID_ENTRY = Symbol('sawValidEntry') | ||
|  | const SAW_NULL_BLOCK = Symbol('sawNullBlock') | ||
|  | const SAW_EOF = Symbol('sawEOF') | ||
|  | const CLOSESTREAM = Symbol('closeStream') | ||
|  | 
 | ||
|  | const noop = _ => true | ||
|  | 
 | ||
|  | module.exports = warner(class Parser extends EE { | ||
|  |   constructor (opt) { | ||
|  |     opt = opt || {} | ||
|  |     super(opt) | ||
|  | 
 | ||
|  |     this.file = opt.file || '' | ||
|  | 
 | ||
|  |     // set to boolean false when an entry starts.  1024 bytes of \0
 | ||
|  |     // is technically a valid tarball, albeit a boring one.
 | ||
|  |     this[SAW_VALID_ENTRY] = null | ||
|  | 
 | ||
|  |     // these BADARCHIVE errors can't be detected early. listen on DONE.
 | ||
|  |     this.on(DONE, _ => { | ||
|  |       if (this[STATE] === 'begin' || this[SAW_VALID_ENTRY] === false) { | ||
|  |         // either less than 1 block of data, or all entries were invalid.
 | ||
|  |         // Either way, probably not even a tarball.
 | ||
|  |         this.warn('TAR_BAD_ARCHIVE', 'Unrecognized archive format') | ||
|  |       } | ||
|  |     }) | ||
|  | 
 | ||
|  |     if (opt.ondone) { | ||
|  |       this.on(DONE, opt.ondone) | ||
|  |     } else { | ||
|  |       this.on(DONE, _ => { | ||
|  |         this.emit('prefinish') | ||
|  |         this.emit('finish') | ||
|  |         this.emit('end') | ||
|  |       }) | ||
|  |     } | ||
|  | 
 | ||
|  |     this.strict = !!opt.strict | ||
|  |     this.maxMetaEntrySize = opt.maxMetaEntrySize || maxMetaEntrySize | ||
|  |     this.filter = typeof opt.filter === 'function' ? opt.filter : noop | ||
|  | 
 | ||
|  |     // have to set this so that streams are ok piping into it
 | ||
|  |     this.writable = true | ||
|  |     this.readable = false | ||
|  | 
 | ||
|  |     this[QUEUE] = new Yallist() | ||
|  |     this[BUFFER] = null | ||
|  |     this[READENTRY] = null | ||
|  |     this[WRITEENTRY] = null | ||
|  |     this[STATE] = 'begin' | ||
|  |     this[META] = '' | ||
|  |     this[EX] = null | ||
|  |     this[GEX] = null | ||
|  |     this[ENDED] = false | ||
|  |     this[UNZIP] = null | ||
|  |     this[ABORTED] = false | ||
|  |     this[SAW_NULL_BLOCK] = false | ||
|  |     this[SAW_EOF] = false | ||
|  | 
 | ||
|  |     this.on('end', () => this[CLOSESTREAM]()) | ||
|  | 
 | ||
|  |     if (typeof opt.onwarn === 'function') { | ||
|  |       this.on('warn', opt.onwarn) | ||
|  |     } | ||
|  |     if (typeof opt.onentry === 'function') { | ||
|  |       this.on('entry', opt.onentry) | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [CONSUMEHEADER] (chunk, position) { | ||
|  |     if (this[SAW_VALID_ENTRY] === null) { | ||
|  |       this[SAW_VALID_ENTRY] = false | ||
|  |     } | ||
|  |     let header | ||
|  |     try { | ||
|  |       header = new Header(chunk, position, this[EX], this[GEX]) | ||
|  |     } catch (er) { | ||
|  |       return this.warn('TAR_ENTRY_INVALID', er) | ||
|  |     } | ||
|  | 
 | ||
|  |     if (header.nullBlock) { | ||
|  |       if (this[SAW_NULL_BLOCK]) { | ||
|  |         this[SAW_EOF] = true | ||
|  |         // ending an archive with no entries.  pointless, but legal.
 | ||
|  |         if (this[STATE] === 'begin') { | ||
|  |           this[STATE] = 'header' | ||
|  |         } | ||
|  |         this[EMIT]('eof') | ||
|  |       } else { | ||
|  |         this[SAW_NULL_BLOCK] = true | ||
|  |         this[EMIT]('nullBlock') | ||
|  |       } | ||
|  |     } else { | ||
|  |       this[SAW_NULL_BLOCK] = false | ||
|  |       if (!header.cksumValid) { | ||
|  |         this.warn('TAR_ENTRY_INVALID', 'checksum failure', { header }) | ||
|  |       } else if (!header.path) { | ||
|  |         this.warn('TAR_ENTRY_INVALID', 'path is required', { header }) | ||
|  |       } else { | ||
|  |         const type = header.type | ||
|  |         if (/^(Symbolic)?Link$/.test(type) && !header.linkpath) { | ||
|  |           this.warn('TAR_ENTRY_INVALID', 'linkpath required', { header }) | ||
|  |         } else if (!/^(Symbolic)?Link$/.test(type) && header.linkpath) { | ||
|  |           this.warn('TAR_ENTRY_INVALID', 'linkpath forbidden', { header }) | ||
|  |         } else { | ||
|  |           const entry = this[WRITEENTRY] = new Entry(header, this[EX], this[GEX]) | ||
|  | 
 | ||
|  |           // we do this for meta & ignored entries as well, because they
 | ||
|  |           // are still valid tar, or else we wouldn't know to ignore them
 | ||
|  |           if (!this[SAW_VALID_ENTRY]) { | ||
|  |             if (entry.remain) { | ||
|  |               // this might be the one!
 | ||
|  |               const onend = () => { | ||
|  |                 if (!entry.invalid) { | ||
|  |                   this[SAW_VALID_ENTRY] = true | ||
|  |                 } | ||
|  |               } | ||
|  |               entry.on('end', onend) | ||
|  |             } else { | ||
|  |               this[SAW_VALID_ENTRY] = true | ||
|  |             } | ||
|  |           } | ||
|  | 
 | ||
|  |           if (entry.meta) { | ||
|  |             if (entry.size > this.maxMetaEntrySize) { | ||
|  |               entry.ignore = true | ||
|  |               this[EMIT]('ignoredEntry', entry) | ||
|  |               this[STATE] = 'ignore' | ||
|  |               entry.resume() | ||
|  |             } else if (entry.size > 0) { | ||
|  |               this[META] = '' | ||
|  |               entry.on('data', c => this[META] += c) | ||
|  |               this[STATE] = 'meta' | ||
|  |             } | ||
|  |           } else { | ||
|  |             this[EX] = null | ||
|  |             entry.ignore = entry.ignore || !this.filter(entry.path, entry) | ||
|  | 
 | ||
|  |             if (entry.ignore) { | ||
|  |               // probably valid, just not something we care about
 | ||
|  |               this[EMIT]('ignoredEntry', entry) | ||
|  |               this[STATE] = entry.remain ? 'ignore' : 'header' | ||
|  |               entry.resume() | ||
|  |             } else { | ||
|  |               if (entry.remain) { | ||
|  |                 this[STATE] = 'body' | ||
|  |               } else { | ||
|  |                 this[STATE] = 'header' | ||
|  |                 entry.end() | ||
|  |               } | ||
|  | 
 | ||
|  |               if (!this[READENTRY]) { | ||
|  |                 this[QUEUE].push(entry) | ||
|  |                 this[NEXTENTRY]() | ||
|  |               } else { | ||
|  |                 this[QUEUE].push(entry) | ||
|  |               } | ||
|  |             } | ||
|  |           } | ||
|  |         } | ||
|  |       } | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [CLOSESTREAM] () { | ||
|  |     nextTick(() => this.emit('close')) | ||
|  |   } | ||
|  | 
 | ||
|  |   [PROCESSENTRY] (entry) { | ||
|  |     let go = true | ||
|  | 
 | ||
|  |     if (!entry) { | ||
|  |       this[READENTRY] = null | ||
|  |       go = false | ||
|  |     } else if (Array.isArray(entry)) { | ||
|  |       this.emit.apply(this, entry) | ||
|  |     } else { | ||
|  |       this[READENTRY] = entry | ||
|  |       this.emit('entry', entry) | ||
|  |       if (!entry.emittedEnd) { | ||
|  |         entry.on('end', _ => this[NEXTENTRY]()) | ||
|  |         go = false | ||
|  |       } | ||
|  |     } | ||
|  | 
 | ||
|  |     return go | ||
|  |   } | ||
|  | 
 | ||
|  |   [NEXTENTRY] () { | ||
|  |     do {} while (this[PROCESSENTRY](this[QUEUE].shift())) | ||
|  | 
 | ||
|  |     if (!this[QUEUE].length) { | ||
|  |       // At this point, there's nothing in the queue, but we may have an
 | ||
|  |       // entry which is being consumed (readEntry).
 | ||
|  |       // If we don't, then we definitely can handle more data.
 | ||
|  |       // If we do, and either it's flowing, or it has never had any data
 | ||
|  |       // written to it, then it needs more.
 | ||
|  |       // The only other possibility is that it has returned false from a
 | ||
|  |       // write() call, so we wait for the next drain to continue.
 | ||
|  |       const re = this[READENTRY] | ||
|  |       const drainNow = !re || re.flowing || re.size === re.remain | ||
|  |       if (drainNow) { | ||
|  |         if (!this[WRITING]) { | ||
|  |           this.emit('drain') | ||
|  |         } | ||
|  |       } else { | ||
|  |         re.once('drain', _ => this.emit('drain')) | ||
|  |       } | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [CONSUMEBODY] (chunk, position) { | ||
|  |     // write up to but no  more than writeEntry.blockRemain
 | ||
|  |     const entry = this[WRITEENTRY] | ||
|  |     const br = entry.blockRemain | ||
|  |     const c = (br >= chunk.length && position === 0) ? chunk | ||
|  |       : chunk.slice(position, position + br) | ||
|  | 
 | ||
|  |     entry.write(c) | ||
|  | 
 | ||
|  |     if (!entry.blockRemain) { | ||
|  |       this[STATE] = 'header' | ||
|  |       this[WRITEENTRY] = null | ||
|  |       entry.end() | ||
|  |     } | ||
|  | 
 | ||
|  |     return c.length | ||
|  |   } | ||
|  | 
 | ||
|  |   [CONSUMEMETA] (chunk, position) { | ||
|  |     const entry = this[WRITEENTRY] | ||
|  |     const ret = this[CONSUMEBODY](chunk, position) | ||
|  | 
 | ||
|  |     // if we finished, then the entry is reset
 | ||
|  |     if (!this[WRITEENTRY]) { | ||
|  |       this[EMITMETA](entry) | ||
|  |     } | ||
|  | 
 | ||
|  |     return ret | ||
|  |   } | ||
|  | 
 | ||
|  |   [EMIT] (ev, data, extra) { | ||
|  |     if (!this[QUEUE].length && !this[READENTRY]) { | ||
|  |       this.emit(ev, data, extra) | ||
|  |     } else { | ||
|  |       this[QUEUE].push([ev, data, extra]) | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [EMITMETA] (entry) { | ||
|  |     this[EMIT]('meta', this[META]) | ||
|  |     switch (entry.type) { | ||
|  |       case 'ExtendedHeader': | ||
|  |       case 'OldExtendedHeader': | ||
|  |         this[EX] = Pax.parse(this[META], this[EX], false) | ||
|  |         break | ||
|  | 
 | ||
|  |       case 'GlobalExtendedHeader': | ||
|  |         this[GEX] = Pax.parse(this[META], this[GEX], true) | ||
|  |         break | ||
|  | 
 | ||
|  |       case 'NextFileHasLongPath': | ||
|  |       case 'OldGnuLongPath': | ||
|  |         this[EX] = this[EX] || Object.create(null) | ||
|  |         this[EX].path = this[META].replace(/\0.*/, '') | ||
|  |         break | ||
|  | 
 | ||
|  |       case 'NextFileHasLongLinkpath': | ||
|  |         this[EX] = this[EX] || Object.create(null) | ||
|  |         this[EX].linkpath = this[META].replace(/\0.*/, '') | ||
|  |         break | ||
|  | 
 | ||
|  |       /* istanbul ignore next */ | ||
|  |       default: throw new Error('unknown meta: ' + entry.type) | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   abort (error) { | ||
|  |     this[ABORTED] = true | ||
|  |     this.emit('abort', error) | ||
|  |     // always throws, even in non-strict mode
 | ||
|  |     this.warn('TAR_ABORT', error, { recoverable: false }) | ||
|  |   } | ||
|  | 
 | ||
|  |   write (chunk) { | ||
|  |     if (this[ABORTED]) { | ||
|  |       return | ||
|  |     } | ||
|  | 
 | ||
|  |     // first write, might be gzipped
 | ||
|  |     if (this[UNZIP] === null && chunk) { | ||
|  |       if (this[BUFFER]) { | ||
|  |         chunk = Buffer.concat([this[BUFFER], chunk]) | ||
|  |         this[BUFFER] = null | ||
|  |       } | ||
|  |       if (chunk.length < gzipHeader.length) { | ||
|  |         this[BUFFER] = chunk | ||
|  |         return true | ||
|  |       } | ||
|  |       for (let i = 0; this[UNZIP] === null && i < gzipHeader.length; i++) { | ||
|  |         if (chunk[i] !== gzipHeader[i]) { | ||
|  |           this[UNZIP] = false | ||
|  |         } | ||
|  |       } | ||
|  |       if (this[UNZIP] === null) { | ||
|  |         const ended = this[ENDED] | ||
|  |         this[ENDED] = false | ||
|  |         this[UNZIP] = new zlib.Unzip() | ||
|  |         this[UNZIP].on('data', chunk => this[CONSUMECHUNK](chunk)) | ||
|  |         this[UNZIP].on('error', er => this.abort(er)) | ||
|  |         this[UNZIP].on('end', _ => { | ||
|  |           this[ENDED] = true | ||
|  |           this[CONSUMECHUNK]() | ||
|  |         }) | ||
|  |         this[WRITING] = true | ||
|  |         const ret = this[UNZIP][ended ? 'end' : 'write'](chunk) | ||
|  |         this[WRITING] = false | ||
|  |         return ret | ||
|  |       } | ||
|  |     } | ||
|  | 
 | ||
|  |     this[WRITING] = true | ||
|  |     if (this[UNZIP]) { | ||
|  |       this[UNZIP].write(chunk) | ||
|  |     } else { | ||
|  |       this[CONSUMECHUNK](chunk) | ||
|  |     } | ||
|  |     this[WRITING] = false | ||
|  | 
 | ||
|  |     // return false if there's a queue, or if the current entry isn't flowing
 | ||
|  |     const ret = | ||
|  |       this[QUEUE].length ? false : | ||
|  |       this[READENTRY] ? this[READENTRY].flowing : | ||
|  |       true | ||
|  | 
 | ||
|  |     // if we have no queue, then that means a clogged READENTRY
 | ||
|  |     if (!ret && !this[QUEUE].length) { | ||
|  |       this[READENTRY].once('drain', _ => this.emit('drain')) | ||
|  |     } | ||
|  | 
 | ||
|  |     return ret | ||
|  |   } | ||
|  | 
 | ||
|  |   [BUFFERCONCAT] (c) { | ||
|  |     if (c && !this[ABORTED]) { | ||
|  |       this[BUFFER] = this[BUFFER] ? Buffer.concat([this[BUFFER], c]) : c | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [MAYBEEND] () { | ||
|  |     if (this[ENDED] && | ||
|  |         !this[EMITTEDEND] && | ||
|  |         !this[ABORTED] && | ||
|  |         !this[CONSUMING]) { | ||
|  |       this[EMITTEDEND] = true | ||
|  |       const entry = this[WRITEENTRY] | ||
|  |       if (entry && entry.blockRemain) { | ||
|  |         // truncated, likely a damaged file
 | ||
|  |         const have = this[BUFFER] ? this[BUFFER].length : 0 | ||
|  |         this.warn('TAR_BAD_ARCHIVE', `Truncated input (needed ${ | ||
|  |           entry.blockRemain} more bytes, only ${have} available)`, { entry })
 | ||
|  |         if (this[BUFFER]) { | ||
|  |           entry.write(this[BUFFER]) | ||
|  |         } | ||
|  |         entry.end() | ||
|  |       } | ||
|  |       this[EMIT](DONE) | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [CONSUMECHUNK] (chunk) { | ||
|  |     if (this[CONSUMING]) { | ||
|  |       this[BUFFERCONCAT](chunk) | ||
|  |     } else if (!chunk && !this[BUFFER]) { | ||
|  |       this[MAYBEEND]() | ||
|  |     } else { | ||
|  |       this[CONSUMING] = true | ||
|  |       if (this[BUFFER]) { | ||
|  |         this[BUFFERCONCAT](chunk) | ||
|  |         const c = this[BUFFER] | ||
|  |         this[BUFFER] = null | ||
|  |         this[CONSUMECHUNKSUB](c) | ||
|  |       } else { | ||
|  |         this[CONSUMECHUNKSUB](chunk) | ||
|  |       } | ||
|  | 
 | ||
|  |       while (this[BUFFER] && | ||
|  |           this[BUFFER].length >= 512 && | ||
|  |           !this[ABORTED] && | ||
|  |           !this[SAW_EOF]) { | ||
|  |         const c = this[BUFFER] | ||
|  |         this[BUFFER] = null | ||
|  |         this[CONSUMECHUNKSUB](c) | ||
|  |       } | ||
|  |       this[CONSUMING] = false | ||
|  |     } | ||
|  | 
 | ||
|  |     if (!this[BUFFER] || this[ENDED]) { | ||
|  |       this[MAYBEEND]() | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   [CONSUMECHUNKSUB] (chunk) { | ||
|  |     // we know that we are in CONSUMING mode, so anything written goes into
 | ||
|  |     // the buffer.  Advance the position and put any remainder in the buffer.
 | ||
|  |     let position = 0 | ||
|  |     const length = chunk.length | ||
|  |     while (position + 512 <= length && !this[ABORTED] && !this[SAW_EOF]) { | ||
|  |       switch (this[STATE]) { | ||
|  |         case 'begin': | ||
|  |         case 'header': | ||
|  |           this[CONSUMEHEADER](chunk, position) | ||
|  |           position += 512 | ||
|  |           break | ||
|  | 
 | ||
|  |         case 'ignore': | ||
|  |         case 'body': | ||
|  |           position += this[CONSUMEBODY](chunk, position) | ||
|  |           break | ||
|  | 
 | ||
|  |         case 'meta': | ||
|  |           position += this[CONSUMEMETA](chunk, position) | ||
|  |           break | ||
|  | 
 | ||
|  |         /* istanbul ignore next */ | ||
|  |         default: | ||
|  |           throw new Error('invalid state: ' + this[STATE]) | ||
|  |       } | ||
|  |     } | ||
|  | 
 | ||
|  |     if (position < length) { | ||
|  |       if (this[BUFFER]) { | ||
|  |         this[BUFFER] = Buffer.concat([chunk.slice(position), this[BUFFER]]) | ||
|  |       } else { | ||
|  |         this[BUFFER] = chunk.slice(position) | ||
|  |       } | ||
|  |     } | ||
|  |   } | ||
|  | 
 | ||
|  |   end (chunk) { | ||
|  |     if (!this[ABORTED]) { | ||
|  |       if (this[UNZIP]) { | ||
|  |         this[UNZIP].end(chunk) | ||
|  |       } else { | ||
|  |         this[ENDED] = true | ||
|  |         this.write(chunk) | ||
|  |       } | ||
|  |     } | ||
|  |   } | ||
|  | }) |