@@ -1290,7 +1290,6 @@ def predict(
1290
1290
frequency_penalty : Optional [float ] = None ,
1291
1291
logit_bias : Optional [Dict [int , float ]] = None ,
1292
1292
seed : Optional [int ] = None ,
1293
- echo : Optional [bool ] = None ,
1294
1293
) -> "MultiCandidateTextGenerationResponse" :
1295
1294
"""Gets model response for a single prompt.
1296
1295
@@ -1329,8 +1328,6 @@ def predict(
1329
1328
the same output with the same seed. If seed is not set, the seed used in decoder will not be
1330
1329
deterministic, thus the generated random noise will not be deterministic. If seed is set, the
1331
1330
generated random noise will be deterministic.
1332
- echo:
1333
- If true, the prompt is echoed in the generated text.
1334
1331
1335
1332
Returns:
1336
1333
A `MultiCandidateTextGenerationResponse` object that contains the text produced by the model.
@@ -1349,7 +1346,6 @@ def predict(
1349
1346
frequency_penalty = frequency_penalty ,
1350
1347
logit_bias = logit_bias ,
1351
1348
seed = seed ,
1352
- echo = echo ,
1353
1349
)
1354
1350
1355
1351
prediction_response = self ._endpoint .predict (
@@ -1383,7 +1379,6 @@ async def predict_async(
1383
1379
frequency_penalty : Optional [float ] = None ,
1384
1380
logit_bias : Optional [Dict [int , float ]] = None ,
1385
1381
seed : Optional [int ] = None ,
1386
- echo : Optional [bool ] = None ,
1387
1382
) -> "MultiCandidateTextGenerationResponse" :
1388
1383
"""Asynchronously gets model response for a single prompt.
1389
1384
@@ -1422,8 +1417,6 @@ async def predict_async(
1422
1417
the same output with the same seed. If seed is not set, the seed used in decoder will not be
1423
1418
deterministic, thus the generated random noise will not be deterministic. If seed is set, the
1424
1419
generated random noise will be deterministic.
1425
- echo:
1426
- If true, the prompt is echoed in the generated text.
1427
1420
1428
1421
Returns:
1429
1422
A `MultiCandidateTextGenerationResponse` object that contains the text produced by the model.
@@ -1442,7 +1435,6 @@ async def predict_async(
1442
1435
frequency_penalty = frequency_penalty ,
1443
1436
logit_bias = logit_bias ,
1444
1437
seed = seed ,
1445
- echo = echo ,
1446
1438
)
1447
1439
1448
1440
prediction_response = await self ._endpoint .predict_async (
@@ -1468,7 +1460,6 @@ def predict_streaming(
1468
1460
frequency_penalty : Optional [float ] = None ,
1469
1461
logit_bias : Optional [Dict [int , float ]] = None ,
1470
1462
seed : Optional [int ] = None ,
1471
- echo : Optional [bool ] = None ,
1472
1463
) -> Iterator [TextGenerationResponse ]:
1473
1464
"""Gets a streaming model response for a single prompt.
1474
1465
@@ -1507,8 +1498,6 @@ def predict_streaming(
1507
1498
the same output with the same seed. If seed is not set, the seed used in decoder will not be
1508
1499
deterministic, thus the generated random noise will not be deterministic. If seed is set, the
1509
1500
generated random noise will be deterministic.
1510
- echo:
1511
- If true, the prompt is echoed in the generated text.
1512
1501
1513
1502
Yields:
1514
1503
A stream of `TextGenerationResponse` objects that contain partial
@@ -1526,7 +1515,6 @@ def predict_streaming(
1526
1515
frequency_penalty = frequency_penalty ,
1527
1516
logit_bias = logit_bias ,
1528
1517
seed = seed ,
1529
- echo = echo ,
1530
1518
)
1531
1519
1532
1520
prediction_service_client = self ._endpoint ._prediction_client
@@ -1558,7 +1546,6 @@ async def predict_streaming_async(
1558
1546
frequency_penalty : Optional [float ] = None ,
1559
1547
logit_bias : Optional [Dict [int , float ]] = None ,
1560
1548
seed : Optional [int ] = None ,
1561
- echo : Optional [bool ] = None ,
1562
1549
) -> AsyncIterator [TextGenerationResponse ]:
1563
1550
"""Asynchronously gets a streaming model response for a single prompt.
1564
1551
@@ -1597,8 +1584,6 @@ async def predict_streaming_async(
1597
1584
the same output with the same seed. If seed is not set, the seed used in decoder will not be
1598
1585
deterministic, thus the generated random noise will not be deterministic. If seed is set, the
1599
1586
generated random noise will be deterministic.
1600
- echo:
1601
- If true, the prompt is echoed in the generated text.
1602
1587
1603
1588
Yields:
1604
1589
A stream of `TextGenerationResponse` objects that contain partial
@@ -1616,7 +1601,6 @@ async def predict_streaming_async(
1616
1601
frequency_penalty = frequency_penalty ,
1617
1602
logit_bias = logit_bias ,
1618
1603
seed = seed ,
1619
- echo = echo ,
1620
1604
)
1621
1605
1622
1606
prediction_service_async_client = self ._endpoint ._prediction_async_client
@@ -1654,7 +1638,6 @@ def _create_text_generation_prediction_request(
1654
1638
frequency_penalty : Optional [float ] = None ,
1655
1639
logit_bias : Optional [Dict [int , int ]] = None ,
1656
1640
seed : Optional [int ] = None ,
1657
- echo : Optional [bool ] = None ,
1658
1641
) -> "_PredictionRequest" :
1659
1642
"""Prepares the text generation request for a single prompt.
1660
1643
@@ -1693,8 +1676,6 @@ def _create_text_generation_prediction_request(
1693
1676
the same output with the same seed. If seed is not set, the seed used in decoder will not be
1694
1677
deterministic, thus the generated random noise will not be deterministic. If seed is set, the
1695
1678
generated random noise will be deterministic.
1696
- echo:
1697
- If true, the prompt is echoed in the generated text.
1698
1679
1699
1680
Returns:
1700
1681
A `_PredictionRequest` object that contains prediction instance and parameters.
@@ -1744,9 +1725,6 @@ def _create_text_generation_prediction_request(
1744
1725
if seed is not None :
1745
1726
prediction_parameters ["seed" ] = seed
1746
1727
1747
- if echo is not None :
1748
- prediction_parameters ["echo" ] = echo
1749
-
1750
1728
return _PredictionRequest (
1751
1729
instance = instance ,
1752
1730
parameters = prediction_parameters ,
0 commit comments