# @llamaindex/env

## 0.1.31

### Patch Changes

- 5977b1f: Add deprecation message

## 0.1.30

### Patch Changes

- 9b2e25a: Use Uint8Array instead of Buffer for file type messages (works with non-NodeJS)

## 0.1.29

### Patch Changes

- beb922b: Fix edge runtime builds by adding missing packages to env package. Make gpt-tokenizer optional for llamaindex to reduce package size.

## 0.1.28

### Patch Changes

- cc50c9c: fix: missing release

## 0.1.27

### Patch Changes

- 1c908fd: Revert previous release (not working with CJS)

## 0.1.26

### Patch Changes

- cb608b5: fix: bundle output incorrect

## 0.1.25

### Patch Changes

- b504303: Allow Node 18 again (throw run-time error if not possible) to make Stackblitz work

## 0.1.24

### Patch Changes

- 8be4589: chore: bump version

## 0.1.23

### Patch Changes

- d2b2722: fix: switch tokenizer in cloudflare

## 0.1.22

### Patch Changes

- 969365c: fix: async local storage on cloudflare worker

## 0.1.21

### Patch Changes

- 90d265c: chore: bump version

## 0.1.20

### Patch Changes

- 4fc001c: chore: bump `@huggingface/transformers`

  Upgrade to v3, please read https://github.com/huggingface/transformers.js/releases/tag/3.0.0 for more information.

## 0.1.19

### Patch Changes

- ad85bd0: - fix agent chat message not saved into the task context when streaming
  - fix async local storage might use `node:async_hook` in edge-light/workerd condition

## 0.1.18

### Patch Changes

- a8d3fa6: fix: exports in package.json

## 0.1.17

### Patch Changes

- 14cc9eb: chore: move multi-model into single sub module

## 0.1.16

### Patch Changes

- fa60fc6: fix(env): no esm shim

## 0.1.15

### Patch Changes

- 4ba2cfe: fix(env): align export APIs

## 0.1.14

### Patch Changes

- ae49ff4: feat: use `gpt-tokenizer`
- a75af83: refactor: move some llm and embedding to single package

## 0.1.13

### Patch Changes

- df441e2: fix: consoleLogger is missing from `@llamaindex/env`

## 0.1.12

### Patch Changes

- b48bcc3: feat: add `load-transformers` event type when loading `@xenova/transformers` module

  This would benefit user who want to customize the transformer env.

## 0.1.11

### Patch Changes

- ac07e3c: fix: replace instanceof check with `.type` check
- 1a6137b: feat: experimental support for browser

  If you see bundler issue in next.js edge runtime, please bump to `next@14` latest version.

- ac07e3c: fix: add `console.warn` when import dual module

## 0.1.10

### Patch Changes

- 4648da6: fix: wrong tiktoken version caused NextJs CL template run fail

## 0.1.9

### Patch Changes

- 58abc57: fix: align version

## 0.1.8

### Patch Changes

- f326ab8: chore: bump version

## 0.1.7

### Patch Changes

- 41fe871: Add support for azure dynamic session tool

## 0.1.6

### Patch Changes

- d4e853c: fix: stronger type declaration
- a94b8ec: fix: jsr release

## 0.1.5

### Patch Changes

- f3b34b4: Use tiktoken instead of tiktoken/lite and disable WASM tiktoken for non-Node environments

## 0.1.4

### Patch Changes

- 56fabbb: Release env changes to tokenizer

## 0.1.3

### Patch Changes

- e072c45: fix: remove non-standard API `pipeline`
- 9e133ac: refactor: remove `defaultFS` from parameters

  We don't accept passing fs in the parameter since it's unnecessary for a determined JS environment.

  This was a polyfill way for the non-Node.js environment, but now we use another way to polyfill APIs.

## 0.1.2

### Patch Changes

- efa326a: chore: update package.json
- efa326a: refactor: remove usage of lodash

## 0.1.1

### Patch Changes

- 5596e31: feat: improve `@llamaindex/env`

## 0.1.0

### Minor Changes

- 5016f21: feat: improve next.js/cloudflare/vite support

## 0.0.7

### Patch Changes

- Add polyfill for pipeline

## 0.0.6

### Patch Changes

- 7a23cc6: feat: improve CallbackManager

## 0.0.5

### Patch Changes

- 5116ad8: fix: compatibility issue with Deno

## 0.0.4

### Patch Changes

- cf87f84: fix: type backward compatibility

## 0.0.3

### Patch Changes

- e8e21a0: build: set files in package.json

## 0.0.2

### Patch Changes

- 7416a87: build: cjs file not found
