1
1
# This file was autogenerated by uv via the following command:
2
- # uv export --frozen --no-hashes --no-emit-project --no-default-groups -- output-file=requirements.txt
2
+ # uv export --frozen --no-hashes --no-emit-project --output-file=requirements.txt
3
3
aiohappyeyeballs == 2.5.0
4
4
# via aiohttp
5
5
aiohttp == 3.11.13
@@ -14,36 +14,72 @@ anyio==4.8.0
14
14
# llama-stack-client
15
15
# openai
16
16
# starlette
17
+ appnope == 0.1.4 ; sys_platform == 'darwin'
18
+ # via ipykernel
19
+ asttokens == 3.0.0
20
+ # via stack-data
17
21
async-timeout == 5.0.1 ; python_full_version < '3.11'
18
22
# via aiohttp
19
23
attrs == 25.1.0
20
24
# via
21
25
# aiohttp
22
26
# jsonschema
23
27
# referencing
28
+ black == 25.1.0
24
29
certifi == 2025.1.31
25
30
# via
26
31
# httpcore
27
32
# httpx
28
33
# requests
34
+ cffi == 1.17.1 ; implementation_name == 'pypy'
35
+ # via pyzmq
36
+ cfgv == 3.4.0
37
+ # via pre-commit
29
38
charset-normalizer == 3.4.1
30
39
# via requests
31
40
click == 8.1.8
32
- # via llama-stack-client
41
+ # via
42
+ # black
43
+ # llama-stack-client
44
+ # uvicorn
33
45
colorama == 0.4.6 ; sys_platform == 'win32'
34
46
# via
35
47
# click
48
+ # ipython
49
+ # pytest
36
50
# tqdm
51
+ comm == 0.2.2
52
+ # via ipykernel
53
+ coverage == 7.6.12
54
+ # via
55
+ # nbval
56
+ # pytest-cov
57
+ debugpy == 1.8.12
58
+ # via ipykernel
59
+ decorator == 5.1.1
60
+ # via ipython
61
+ distlib == 0.3.9
62
+ # via virtualenv
37
63
distro == 1.9.0
38
64
# via
39
65
# llama-stack-client
40
66
# openai
41
67
ecdsa == 0.19.1
42
68
# via python-jose
43
69
exceptiongroup == 1.2.2 ; python_full_version < '3.11'
44
- # via anyio
70
+ # via
71
+ # anyio
72
+ # ipython
73
+ # pytest
74
+ executing == 2.2.0
75
+ # via stack-data
76
+ fastapi == 0.115.8
77
+ fastjsonschema == 2.21.1
78
+ # via nbformat
45
79
filelock == 3.17.0
46
- # via huggingface-hub
80
+ # via
81
+ # huggingface-hub
82
+ # virtualenv
47
83
fire == 0.7.0
48
84
# via llama-stack
49
85
frozenlist == 1.5.0
@@ -56,6 +92,7 @@ h11==0.16.0
56
92
# via
57
93
# httpcore
58
94
# llama-stack
95
+ # uvicorn
59
96
httpcore == 1.0.9
60
97
# via httpx
61
98
httpx == 0.28.1
@@ -65,77 +102,170 @@ httpx==0.28.1
65
102
# openai
66
103
huggingface-hub == 0.29.0
67
104
# via llama-stack
105
+ identify == 2.6.7
106
+ # via pre-commit
68
107
idna == 3.10
69
108
# via
70
109
# anyio
71
110
# httpx
72
111
# requests
73
112
# yarl
113
+ iniconfig == 2.0.0
114
+ # via pytest
115
+ ipykernel == 6.29.5
116
+ # via nbval
117
+ ipython == 8.32.0
118
+ # via ipykernel
119
+ jedi == 0.19.2
120
+ # via ipython
74
121
jinja2 == 3.1.6
75
- # via llama-stack
122
+ # via
123
+ # llama-stack
124
+ # pytest-html
76
125
jiter == 0.8.2
77
126
# via openai
78
127
jsonschema == 4.23.0
79
- # via llama-stack
128
+ # via
129
+ # llama-stack
130
+ # nbformat
80
131
jsonschema-specifications == 2024.10.1
81
132
# via jsonschema
133
+ jupyter-client == 8.6.3
134
+ # via
135
+ # ipykernel
136
+ # nbval
137
+ jupyter-core == 5.7.2
138
+ # via
139
+ # ipykernel
140
+ # jupyter-client
141
+ # nbformat
82
142
llama-stack-client == 0.2.8
83
143
# via llama-stack
84
144
markdown-it-py == 3.0.0
85
145
# via rich
86
146
markupsafe == 3.0.2
87
147
# via jinja2
148
+ matplotlib-inline == 0.1.7
149
+ # via
150
+ # ipykernel
151
+ # ipython
88
152
mdurl == 0.1.2
89
153
# via markdown-it-py
90
154
multidict == 6.1.0
91
155
# via
92
156
# aiohttp
93
157
# yarl
158
+ mypy-extensions == 1.0.0
159
+ # via black
160
+ nbformat == 5.10.4
161
+ # via nbval
162
+ nbval == 0.11.0
163
+ nest-asyncio == 1.6.0
164
+ # via ipykernel
165
+ nodeenv == 1.9.1
166
+ # via pre-commit
94
167
numpy == 2.2.3
95
168
# via pandas
96
169
openai == 1.71.0
97
170
# via llama-stack
98
171
packaging == 24.2
99
- # via huggingface-hub
172
+ # via
173
+ # black
174
+ # huggingface-hub
175
+ # ipykernel
176
+ # pytest
100
177
pandas == 2.2.3
101
178
# via llama-stack-client
179
+ parso == 0.8.4
180
+ # via jedi
181
+ pathspec == 0.12.1
182
+ # via black
183
+ pexpect == 4.9.0 ; sys_platform != 'emscripten' and sys_platform != 'win32'
184
+ # via ipython
102
185
pillow == 11.1.0
103
186
# via llama-stack
187
+ platformdirs == 4.3.6
188
+ # via
189
+ # black
190
+ # jupyter-core
191
+ # virtualenv
192
+ pluggy == 1.5.0
193
+ # via pytest
194
+ pre-commit == 4.1.0
104
195
prompt-toolkit == 3.0.50
105
196
# via
197
+ # ipython
106
198
# llama-stack
107
199
# llama-stack-client
108
200
propcache == 0.3.0
109
201
# via
110
202
# aiohttp
111
203
# yarl
204
+ psutil == 7.0.0
205
+ # via ipykernel
206
+ ptyprocess == 0.7.0 ; sys_platform != 'emscripten' and sys_platform != 'win32'
207
+ # via pexpect
208
+ pure-eval == 0.2.3
209
+ # via stack-data
112
210
pyaml == 25.1.0
113
211
# via llama-stack-client
114
212
pyasn1 == 0.4.8
115
213
# via
116
214
# python-jose
117
215
# rsa
216
+ pycparser == 2.22 ; implementation_name == 'pypy'
217
+ # via cffi
118
218
pydantic == 2.10.6
119
219
# via
220
+ # fastapi
120
221
# llama-stack
121
222
# llama-stack-client
122
223
# openai
123
224
pydantic-core == 2.27.2
124
225
# via pydantic
125
226
pygments == 2.19.1
126
- # via rich
227
+ # via
228
+ # ipython
229
+ # rich
230
+ pytest == 8.3.4
231
+ # via
232
+ # nbval
233
+ # pytest-asyncio
234
+ # pytest-cov
235
+ # pytest-html
236
+ # pytest-json-report
237
+ # pytest-metadata
238
+ # pytest-timeout
239
+ pytest-asyncio == 0.25.3
240
+ pytest-cov == 6.0.0
241
+ pytest-html == 4.1.1
242
+ pytest-json-report == 1.5.0
243
+ pytest-metadata == 3.1.1
244
+ # via
245
+ # pytest-html
246
+ # pytest-json-report
247
+ pytest-timeout == 2.4.0
127
248
python-dateutil == 2.9.0.post0
128
- # via pandas
249
+ # via
250
+ # jupyter-client
251
+ # pandas
129
252
python-dotenv == 1.0.1
130
253
# via llama-stack
131
254
python-jose == 3.4.0
132
255
# via llama-stack
133
256
pytz == 2025.1
134
257
# via pandas
258
+ pywin32 == 308 ; platform_python_implementation != 'PyPy' and sys_platform == 'win32'
259
+ # via jupyter-core
135
260
pyyaml == 6.0.2
136
261
# via
137
262
# huggingface-hub
263
+ # pre-commit
138
264
# pyaml
265
+ pyzmq == 26.2.1
266
+ # via
267
+ # ipykernel
268
+ # jupyter-client
139
269
referencing == 0.36.2
140
270
# via
141
271
# jsonschema
@@ -157,6 +287,10 @@ rpds-py==0.22.3
157
287
# referencing
158
288
rsa == 4.9
159
289
# via python-jose
290
+ ruamel-yaml == 0.18.10
291
+ ruamel-yaml-clib == 0.2.12 ; python_full_version < '3.13' and platform_python_implementation == 'CPython'
292
+ # via ruamel-yaml
293
+ ruff == 0.9.6
160
294
setuptools == 80.8.0
161
295
# via llama-stack
162
296
six == 1.17.0
@@ -168,35 +302,68 @@ sniffio==1.3.1
168
302
# anyio
169
303
# llama-stack-client
170
304
# openai
305
+ stack-data == 0.6.3
306
+ # via ipython
171
307
starlette == 0.45.3
172
- # via llama-stack
308
+ # via
309
+ # fastapi
310
+ # llama-stack
173
311
termcolor == 2.5.0
174
312
# via
175
313
# fire
176
314
# llama-stack
177
315
# llama-stack-client
178
316
tiktoken == 0.9.0
179
317
# via llama-stack
318
+ tomli == 2.2.1 ; python_full_version < = '3.11'
319
+ # via
320
+ # black
321
+ # coverage
322
+ # pytest
323
+ tornado == 6.4.2
324
+ # via
325
+ # ipykernel
326
+ # jupyter-client
180
327
tqdm == 4.67.1
181
328
# via
182
329
# huggingface-hub
183
330
# llama-stack-client
184
331
# openai
332
+ traitlets == 5.14.3
333
+ # via
334
+ # comm
335
+ # ipykernel
336
+ # ipython
337
+ # jupyter-client
338
+ # jupyter-core
339
+ # matplotlib-inline
340
+ # nbformat
341
+ types-requests == 2.32.0.20241016
342
+ types-setuptools == 75.8.0.20250210
185
343
typing-extensions == 4.12.2
186
344
# via
187
345
# anyio
346
+ # black
347
+ # fastapi
188
348
# huggingface-hub
349
+ # ipython
189
350
# llama-stack-client
190
351
# multidict
191
352
# openai
192
353
# pydantic
193
354
# pydantic-core
194
355
# referencing
195
356
# rich
357
+ # uvicorn
196
358
tzdata == 2025.1
197
359
# via pandas
198
360
urllib3 == 2.3.0
199
- # via requests
361
+ # via
362
+ # requests
363
+ # types-requests
364
+ uvicorn == 0.34.0
365
+ virtualenv == 20.29.2
366
+ # via pre-commit
200
367
wcwidth == 0.2.13
201
368
# via prompt-toolkit
202
369
yarl == 1.18.3
0 commit comments