Skip to content

Commit 8e79ef9

Browse files
build: Bump version to 0.2.9
1 parent b53dda9 commit 8e79ef9

File tree

3 files changed

+182
-15
lines changed

3 files changed

+182
-15
lines changed

llama_stack/ui/package.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
"@radix-ui/react-tooltip": "^1.2.6",
2020
"class-variance-authority": "^0.7.1",
2121
"clsx": "^2.1.1",
22-
"llama-stack-client": "0.2.8",
22+
"llama-stack-client": "0.2.9",
2323
"lucide-react": "^0.510.0",
2424
"next": "15.3.2",
2525
"next-themes": "^0.4.6",

pyproject.toml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
44

55
[project]
66
name = "llama_stack"
7-
version = "0.2.9rc1"
7+
version = "0.2.9"
88
authors = [{ name = "Meta Llama", email = "[email protected]" }]
99
description = "Llama Stack"
1010
readme = "README.md"
@@ -27,7 +27,7 @@ dependencies = [
2727
"huggingface-hub",
2828
"jinja2>=3.1.6",
2929
"jsonschema",
30-
"llama-stack-client>=0.2.9rc1",
30+
"llama-stack-client>=0.2.9",
3131
"openai>=1.66",
3232
"prompt-toolkit",
3333
"python-dotenv",
@@ -47,7 +47,7 @@ dependencies = [
4747
ui = [
4848
"streamlit",
4949
"pandas",
50-
"llama-stack-client>=0.2.9rc1",
50+
"llama-stack-client>=0.2.9",
5151
"streamlit-option-menu",
5252
]
5353

requirements.txt

Lines changed: 178 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# This file was autogenerated by uv via the following command:
2-
# uv export --frozen --no-hashes --no-emit-project --no-default-groups --output-file=requirements.txt
2+
# uv export --frozen --no-hashes --no-emit-project --output-file=requirements.txt
33
aiohappyeyeballs==2.5.0
44
# via aiohttp
55
aiohttp==3.11.13
@@ -14,36 +14,72 @@ anyio==4.8.0
1414
# llama-stack-client
1515
# openai
1616
# starlette
17+
appnope==0.1.4 ; sys_platform == 'darwin'
18+
# via ipykernel
19+
asttokens==3.0.0
20+
# via stack-data
1721
async-timeout==5.0.1 ; python_full_version < '3.11'
1822
# via aiohttp
1923
attrs==25.1.0
2024
# via
2125
# aiohttp
2226
# jsonschema
2327
# referencing
28+
black==25.1.0
2429
certifi==2025.1.31
2530
# via
2631
# httpcore
2732
# httpx
2833
# requests
34+
cffi==1.17.1 ; implementation_name == 'pypy'
35+
# via pyzmq
36+
cfgv==3.4.0
37+
# via pre-commit
2938
charset-normalizer==3.4.1
3039
# via requests
3140
click==8.1.8
32-
# via llama-stack-client
41+
# via
42+
# black
43+
# llama-stack-client
44+
# uvicorn
3345
colorama==0.4.6 ; sys_platform == 'win32'
3446
# via
3547
# click
48+
# ipython
49+
# pytest
3650
# tqdm
51+
comm==0.2.2
52+
# via ipykernel
53+
coverage==7.6.12
54+
# via
55+
# nbval
56+
# pytest-cov
57+
debugpy==1.8.12
58+
# via ipykernel
59+
decorator==5.1.1
60+
# via ipython
61+
distlib==0.3.9
62+
# via virtualenv
3763
distro==1.9.0
3864
# via
3965
# llama-stack-client
4066
# openai
4167
ecdsa==0.19.1
4268
# via python-jose
4369
exceptiongroup==1.2.2 ; python_full_version < '3.11'
44-
# via anyio
70+
# via
71+
# anyio
72+
# ipython
73+
# pytest
74+
executing==2.2.0
75+
# via stack-data
76+
fastapi==0.115.8
77+
fastjsonschema==2.21.1
78+
# via nbformat
4579
filelock==3.17.0
46-
# via huggingface-hub
80+
# via
81+
# huggingface-hub
82+
# virtualenv
4783
fire==0.7.0
4884
# via llama-stack
4985
frozenlist==1.5.0
@@ -56,6 +92,7 @@ h11==0.16.0
5692
# via
5793
# httpcore
5894
# llama-stack
95+
# uvicorn
5996
httpcore==1.0.9
6097
# via httpx
6198
httpx==0.28.1
@@ -65,77 +102,170 @@ httpx==0.28.1
65102
# openai
66103
huggingface-hub==0.29.0
67104
# via llama-stack
105+
identify==2.6.7
106+
# via pre-commit
68107
idna==3.10
69108
# via
70109
# anyio
71110
# httpx
72111
# requests
73112
# yarl
113+
iniconfig==2.0.0
114+
# via pytest
115+
ipykernel==6.29.5
116+
# via nbval
117+
ipython==8.32.0
118+
# via ipykernel
119+
jedi==0.19.2
120+
# via ipython
74121
jinja2==3.1.6
75-
# via llama-stack
122+
# via
123+
# llama-stack
124+
# pytest-html
76125
jiter==0.8.2
77126
# via openai
78127
jsonschema==4.23.0
79-
# via llama-stack
128+
# via
129+
# llama-stack
130+
# nbformat
80131
jsonschema-specifications==2024.10.1
81132
# via jsonschema
133+
jupyter-client==8.6.3
134+
# via
135+
# ipykernel
136+
# nbval
137+
jupyter-core==5.7.2
138+
# via
139+
# ipykernel
140+
# jupyter-client
141+
# nbformat
82142
llama-stack-client==0.2.8
83143
# via llama-stack
84144
markdown-it-py==3.0.0
85145
# via rich
86146
markupsafe==3.0.2
87147
# via jinja2
148+
matplotlib-inline==0.1.7
149+
# via
150+
# ipykernel
151+
# ipython
88152
mdurl==0.1.2
89153
# via markdown-it-py
90154
multidict==6.1.0
91155
# via
92156
# aiohttp
93157
# yarl
158+
mypy-extensions==1.0.0
159+
# via black
160+
nbformat==5.10.4
161+
# via nbval
162+
nbval==0.11.0
163+
nest-asyncio==1.6.0
164+
# via ipykernel
165+
nodeenv==1.9.1
166+
# via pre-commit
94167
numpy==2.2.3
95168
# via pandas
96169
openai==1.71.0
97170
# via llama-stack
98171
packaging==24.2
99-
# via huggingface-hub
172+
# via
173+
# black
174+
# huggingface-hub
175+
# ipykernel
176+
# pytest
100177
pandas==2.2.3
101178
# via llama-stack-client
179+
parso==0.8.4
180+
# via jedi
181+
pathspec==0.12.1
182+
# via black
183+
pexpect==4.9.0 ; sys_platform != 'emscripten' and sys_platform != 'win32'
184+
# via ipython
102185
pillow==11.1.0
103186
# via llama-stack
187+
platformdirs==4.3.6
188+
# via
189+
# black
190+
# jupyter-core
191+
# virtualenv
192+
pluggy==1.5.0
193+
# via pytest
194+
pre-commit==4.1.0
104195
prompt-toolkit==3.0.50
105196
# via
197+
# ipython
106198
# llama-stack
107199
# llama-stack-client
108200
propcache==0.3.0
109201
# via
110202
# aiohttp
111203
# yarl
204+
psutil==7.0.0
205+
# via ipykernel
206+
ptyprocess==0.7.0 ; sys_platform != 'emscripten' and sys_platform != 'win32'
207+
# via pexpect
208+
pure-eval==0.2.3
209+
# via stack-data
112210
pyaml==25.1.0
113211
# via llama-stack-client
114212
pyasn1==0.4.8
115213
# via
116214
# python-jose
117215
# rsa
216+
pycparser==2.22 ; implementation_name == 'pypy'
217+
# via cffi
118218
pydantic==2.10.6
119219
# via
220+
# fastapi
120221
# llama-stack
121222
# llama-stack-client
122223
# openai
123224
pydantic-core==2.27.2
124225
# via pydantic
125226
pygments==2.19.1
126-
# via rich
227+
# via
228+
# ipython
229+
# rich
230+
pytest==8.3.4
231+
# via
232+
# nbval
233+
# pytest-asyncio
234+
# pytest-cov
235+
# pytest-html
236+
# pytest-json-report
237+
# pytest-metadata
238+
# pytest-timeout
239+
pytest-asyncio==0.25.3
240+
pytest-cov==6.0.0
241+
pytest-html==4.1.1
242+
pytest-json-report==1.5.0
243+
pytest-metadata==3.1.1
244+
# via
245+
# pytest-html
246+
# pytest-json-report
247+
pytest-timeout==2.4.0
127248
python-dateutil==2.9.0.post0
128-
# via pandas
249+
# via
250+
# jupyter-client
251+
# pandas
129252
python-dotenv==1.0.1
130253
# via llama-stack
131254
python-jose==3.4.0
132255
# via llama-stack
133256
pytz==2025.1
134257
# via pandas
258+
pywin32==308 ; platform_python_implementation != 'PyPy' and sys_platform == 'win32'
259+
# via jupyter-core
135260
pyyaml==6.0.2
136261
# via
137262
# huggingface-hub
263+
# pre-commit
138264
# pyaml
265+
pyzmq==26.2.1
266+
# via
267+
# ipykernel
268+
# jupyter-client
139269
referencing==0.36.2
140270
# via
141271
# jsonschema
@@ -157,6 +287,10 @@ rpds-py==0.22.3
157287
# referencing
158288
rsa==4.9
159289
# via python-jose
290+
ruamel-yaml==0.18.10
291+
ruamel-yaml-clib==0.2.12 ; python_full_version < '3.13' and platform_python_implementation == 'CPython'
292+
# via ruamel-yaml
293+
ruff==0.9.6
160294
setuptools==80.8.0
161295
# via llama-stack
162296
six==1.17.0
@@ -168,35 +302,68 @@ sniffio==1.3.1
168302
# anyio
169303
# llama-stack-client
170304
# openai
305+
stack-data==0.6.3
306+
# via ipython
171307
starlette==0.45.3
172-
# via llama-stack
308+
# via
309+
# fastapi
310+
# llama-stack
173311
termcolor==2.5.0
174312
# via
175313
# fire
176314
# llama-stack
177315
# llama-stack-client
178316
tiktoken==0.9.0
179317
# via llama-stack
318+
tomli==2.2.1 ; python_full_version <= '3.11'
319+
# via
320+
# black
321+
# coverage
322+
# pytest
323+
tornado==6.4.2
324+
# via
325+
# ipykernel
326+
# jupyter-client
180327
tqdm==4.67.1
181328
# via
182329
# huggingface-hub
183330
# llama-stack-client
184331
# openai
332+
traitlets==5.14.3
333+
# via
334+
# comm
335+
# ipykernel
336+
# ipython
337+
# jupyter-client
338+
# jupyter-core
339+
# matplotlib-inline
340+
# nbformat
341+
types-requests==2.32.0.20241016
342+
types-setuptools==75.8.0.20250210
185343
typing-extensions==4.12.2
186344
# via
187345
# anyio
346+
# black
347+
# fastapi
188348
# huggingface-hub
349+
# ipython
189350
# llama-stack-client
190351
# multidict
191352
# openai
192353
# pydantic
193354
# pydantic-core
194355
# referencing
195356
# rich
357+
# uvicorn
196358
tzdata==2025.1
197359
# via pandas
198360
urllib3==2.3.0
199-
# via requests
361+
# via
362+
# requests
363+
# types-requests
364+
uvicorn==0.34.0
365+
virtualenv==20.29.2
366+
# via pre-commit
200367
wcwidth==0.2.13
201368
# via prompt-toolkit
202369
yarl==1.18.3

0 commit comments

Comments
 (0)