webllm-llama-3.2-chat / package.json
neetnestor's picture
feat: clone with llama 3.2 models
2819fb4
raw
history blame
759 Bytes
{
"name": "webllm-simple-chat",
"version": "1.0.0",
"main": "dist/index.js",
"scripts": {
"build": "webpack --config webpack.config.js"
},
"dependencies": {
"@babel/core": "^7.13.15",
"@babel/preset-env": "^7.13.15",
"@mlc-ai/web-llm": "^0.2.67",
"katex": "^0.16.11",
"rehype-highlight": "^7.0.0",
"rehype-katex": "^7.0.0",
"rehype-parse": "^9.0.0",
"rehype-stringify": "^10.0.0",
"remark-breaks": "^4.0.0",
"remark-frontmatter": "^5.0.0",
"remark-gfm": "^4.0.0",
"remark-math": "^6.0.0",
"remark-parse": "^11.0.0",
"remark-rehype": "^11.1.0",
"unified": "^11.0.5"
},
"devDependencies": {
"babel-loader": "^8.2.2",
"webpack": "^5.28.0",
"webpack-cli": "^4.6.0"
}
}