| // META: global=window,worker,shadowrealm |
| // META: script=third_party/pako/pako_inflate.min.js |
| // META: script=resources/concatenate-stream.js |
| // META: script=resources/decompress.js |
| // META: script=resources/formats.js |
| // META: timeout=long |
| |
| 'use strict'; |
| |
| // This test verifies that a large flush output will not truncate the |
| // final results. |
| |
| async function compressData(chunk, format) { |
| const cs = new CompressionStream(format); |
| const writer = cs.writable.getWriter(); |
| writer.write(chunk); |
| writer.close(); |
| return await concatenateStream(cs.readable); |
| } |
| |
| // JSON-encoded array of 10 thousands numbers ("[0,1,2,...]"). This produces 48_891 bytes of data. |
| const fullData = new TextEncoder().encode(JSON.stringify(Array.from({ length: 10_000 }, (_, i) => i))); |
| const data = fullData.subarray(0, 35_579); |
| const expectedValue = data; |
| |
| for (const format of formats) { |
| promise_test(async t => { |
| const compressedData = await compressData(data, format); |
| const decompressedData = await decompressDataOrPako(compressedData, format); |
| // check that we got the same result as our original string |
| assert_array_equals(decompressedData, expectedValue, 'value should match'); |
| }, `${format} compression with large flush output`); |
| } |