Skip to content

Commit de22c0e

Browse files
chore: update SDK settings
1 parent 7e697f6 commit de22c0e

19 files changed

+204
-42
lines changed

.github/workflows/ci.yml

Lines changed: 19 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,7 @@ jobs:
1717
timeout-minutes: 10
1818
name: lint
1919
runs-on: ${{ github.repository == 'stainless-sdks/llama-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }}
20+
if: github.event_name == 'push' || github.event.pull_request.head.repo.fork
2021
steps:
2122
- uses: actions/checkout@v4
2223

@@ -34,17 +35,31 @@ jobs:
3435
- name: Run lints
3536
run: ./scripts/lint
3637

37-
upload:
38-
if: github.repository == 'stainless-sdks/llama-api-python'
38+
build:
39+
if: github.repository == 'stainless-sdks/llama-api-python' && (github.event_name == 'push' || github.event.pull_request.head.repo.fork)
3940
timeout-minutes: 10
40-
name: upload
41+
name: build
4142
permissions:
4243
contents: read
4344
id-token: write
4445
runs-on: depot-ubuntu-24.04
4546
steps:
4647
- uses: actions/checkout@v4
4748

49+
- name: Install Rye
50+
run: |
51+
curl -sSf https://rye.astral.sh/get | bash
52+
echo "$HOME/.rye/shims" >> $GITHUB_PATH
53+
env:
54+
RYE_VERSION: '0.44.0'
55+
RYE_INSTALL_OPTION: '--yes'
56+
57+
- name: Install dependencies
58+
run: rye sync --all-features
59+
60+
- name: Run build
61+
run: rye build
62+
4863
- name: Get GitHub OIDC Token
4964
id: github-oidc
5065
uses: actions/github-script@v6
@@ -62,6 +77,7 @@ jobs:
6277
timeout-minutes: 10
6378
name: test
6479
runs-on: ${{ github.repository == 'stainless-sdks/llama-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }}
80+
if: github.event_name == 'push' || github.event.pull_request.head.repo.fork
6581
steps:
6682
- uses: actions/checkout@v4
6783

.github/workflows/publish-pypi.yml

Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
# This workflow is triggered when a GitHub release is created.
2+
# It can also be run manually to re-publish to PyPI in case it failed for some reason.
3+
# You can run this workflow by navigating to https://www.github.com/meta-llama/llama-api-python/actions/workflows/publish-pypi.yml
4+
name: Publish PyPI
5+
on:
6+
workflow_dispatch:
7+
8+
release:
9+
types: [published]
10+
11+
jobs:
12+
publish:
13+
name: publish
14+
runs-on: ubuntu-latest
15+
16+
steps:
17+
- uses: actions/checkout@v4
18+
19+
- name: Install Rye
20+
run: |
21+
curl -sSf https://rye.astral.sh/get | bash
22+
echo "$HOME/.rye/shims" >> $GITHUB_PATH
23+
env:
24+
RYE_VERSION: '0.44.0'
25+
RYE_INSTALL_OPTION: '--yes'
26+
27+
- name: Publish to PyPI
28+
run: |
29+
bash ./bin/publish-pypi
30+
env:
31+
PYPI_TOKEN: ${{ secrets.LLAMA_API_CLIENT_PYPI_TOKEN || secrets.PYPI_TOKEN }}

.github/workflows/release-doctor.yml

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
name: Release Doctor
2+
on:
3+
pull_request:
4+
branches:
5+
- main
6+
workflow_dispatch:
7+
8+
jobs:
9+
release_doctor:
10+
name: release doctor
11+
runs-on: ubuntu-latest
12+
if: github.repository == 'meta-llama/llama-api-python' && (github.event_name == 'push' || github.event_name == 'workflow_dispatch' || startsWith(github.head_ref, 'release-please') || github.head_ref == 'next')
13+
14+
steps:
15+
- uses: actions/checkout@v4
16+
17+
- name: Check release environment
18+
run: |
19+
bash ./bin/check-release-environment
20+
env:
21+
PYPI_TOKEN: ${{ secrets.LLAMA_API_CLIENT_PYPI_TOKEN || secrets.PYPI_TOKEN }}

.release-please-manifest.json

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
{
2+
".": "0.1.2"
3+
}

.stats.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
configured_endpoints: 4
22
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/meta%2Fllama-api-bfa0267b010dcc4b39e62dfbd698ac6f9421f3212c44b3408b9b154bd6c67a8b.yml
33
openapi_spec_hash: 7f424537bc7ea7638e3934ef721b8d71
4-
config_hash: 3ae62c8625d97ed8a867ab369f591724
4+
config_hash: 675e5003e5c68b62e813de9c86b50e35

CONTRIBUTING.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ If you’d like to use the repository from source, you can either install from g
6262
To install via git:
6363

6464
```sh
65-
$ pip install git+ssh://git@github.com/stainless-sdks/llama-api-python.git
65+
$ pip install git+ssh://git@github.com/meta-llama/llama-api-python.git
6666
```
6767

6868
Alternatively, you can build from source and install the wheel file:
@@ -120,7 +120,7 @@ the changes aren't made through the automated pipeline, you may want to make rel
120120

121121
### Publish with a GitHub workflow
122122

123-
You can release to package managers by using [the `Publish PyPI` GitHub action](https://www.github.com/stainless-sdks/llama-api-python/actions/workflows/publish-pypi.yml). This requires a setup organization or repository secret to be set up.
123+
You can release to package managers by using [the `Publish PyPI` GitHub action](https://www.github.com/meta-llama/llama-api-python/actions/workflows/publish-pypi.yml). This requires a setup organization or repository secret to be set up.
124124

125125
### Publish manually
126126

README.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -15,12 +15,12 @@ The REST API documentation can be found on [llama.developer.meta.com](https://ll
1515
## Installation
1616

1717
```sh
18-
# install from this staging repo
19-
pip install git+ssh://git@github.com/stainless-sdks/llama-api-python.git
18+
# install from the production repo
19+
pip install git+ssh://git@github.com/meta-llama/llama-api-python.git
2020
```
2121

2222
> [!NOTE]
23-
> Once this package is [published to PyPI](https://www.stainless.com/docs/guides/publish), this will become: `pip install --pre llama_api_client`
23+
> Once this package is [published to PyPI](https://www.stainless.com/docs/guides/publish), this will become: `pip install llama_api_client`
2424
2525
## Usage
2626

@@ -90,8 +90,8 @@ By default, the async client uses `httpx` for HTTP requests. However, for improv
9090
You can enable this by installing `aiohttp`:
9191

9292
```sh
93-
# install from this staging repo
94-
pip install 'llama_api_client[aiohttp] @ git+ssh://git@github.com/stainless-sdks/llama-api-python.git'
93+
# install from the production repo
94+
pip install 'llama_api_client[aiohttp] @ git+ssh://git@github.com/meta-llama/llama-api-python.git'
9595
```
9696

9797
Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`:
@@ -336,9 +336,9 @@ completion = response.parse() # get the object that `chat.completions.create()`
336336
print(completion.id)
337337
```
338338

339-
These methods return an [`APIResponse`](https://github.com/stainless-sdks/llama-api-python/tree/main/src/llama_api_client/_response.py) object.
339+
These methods return an [`APIResponse`](https://github.com/meta-llama/llama-api-python/tree/main/src/llama_api_client/_response.py) object.
340340

341-
The async client returns an [`AsyncAPIResponse`](https://github.com/stainless-sdks/llama-api-python/tree/main/src/llama_api_client/_response.py) with the same structure, the only difference being `await`able methods for reading the response content.
341+
The async client returns an [`AsyncAPIResponse`](https://github.com/meta-llama/llama-api-python/tree/main/src/llama_api_client/_response.py) with the same structure, the only difference being `await`able methods for reading the response content.
342342

343343
#### `.with_streaming_response`
344344

@@ -450,7 +450,7 @@ This package generally follows [SemVer](https://semver.org/spec/v2.0.0.html) con
450450

451451
We take backwards-compatibility seriously and work hard to ensure you can rely on a smooth upgrade experience.
452452

453-
We are keen for your feedback; please open an [issue](https://www.github.com/stainless-sdks/llama-api-python/issues) with questions, bugs, or suggestions.
453+
We are keen for your feedback; please open an [issue](https://www.github.com/meta-llama/llama-api-python/issues) with questions, bugs, or suggestions.
454454

455455
### Determining the installed version
456456

bin/check-release-environment

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
#!/usr/bin/env bash
2+
3+
errors=()
4+
5+
if [ -z "${PYPI_TOKEN}" ]; then
6+
errors+=("The PYPI_TOKEN secret has not been set. Please set it in either this repository's secrets or your organization secrets.")
7+
fi
8+
9+
lenErrors=${#errors[@]}
10+
11+
if [[ lenErrors -gt 0 ]]; then
12+
echo -e "Found the following errors in the release environment:\n"
13+
14+
for error in "${errors[@]}"; do
15+
echo -e "- $error\n"
16+
done
17+
18+
exit 1
19+
fi
20+
21+
echo "The environment is ready to push releases!"

pyproject.toml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[project]
22
name = "llama_api_client"
3-
version = "0.0.1-alpha.0"
3+
version = "0.1.2"
44
description = "The official Python library for the llama-api-client API"
55
dynamic = ["readme"]
66
license = "MIT"
@@ -34,8 +34,8 @@ classifiers = [
3434
]
3535

3636
[project.urls]
37-
Homepage = "https://github.com/stainless-sdks/llama-api-python"
38-
Repository = "https://github.com/stainless-sdks/llama-api-python"
37+
Homepage = "https://github.com/meta-llama/llama-api-python"
38+
Repository = "https://github.com/meta-llama/llama-api-python"
3939

4040
[project.optional-dependencies]
4141
aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.6"]
@@ -124,7 +124,7 @@ path = "README.md"
124124
[[tool.hatch.metadata.hooks.fancy-pypi-readme.substitutions]]
125125
# replace relative links with absolute links
126126
pattern = '\[(.+?)\]\(((?!https?://)\S+?)\)'
127-
replacement = '[\1](https://github.com/stainless-sdks/llama-api-python/tree/main/\g<2>)'
127+
replacement = '[\1](https://github.com/meta-llama/llama-api-python/tree/main/\g<2>)'
128128

129129
[tool.pytest.ini_options]
130130
testpaths = ["tests"]

release-please-config.json

Lines changed: 66 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,66 @@
1+
{
2+
"packages": {
3+
".": {}
4+
},
5+
"$schema": "https://raw.githubusercontent.com/stainless-api/release-please/main/schemas/config.json",
6+
"include-v-in-tag": true,
7+
"include-component-in-tag": false,
8+
"versioning": "prerelease",
9+
"prerelease": true,
10+
"bump-minor-pre-major": true,
11+
"bump-patch-for-minor-pre-major": false,
12+
"pull-request-header": "Automated Release PR",
13+
"pull-request-title-pattern": "release: ${version}",
14+
"changelog-sections": [
15+
{
16+
"type": "feat",
17+
"section": "Features"
18+
},
19+
{
20+
"type": "fix",
21+
"section": "Bug Fixes"
22+
},
23+
{
24+
"type": "perf",
25+
"section": "Performance Improvements"
26+
},
27+
{
28+
"type": "revert",
29+
"section": "Reverts"
30+
},
31+
{
32+
"type": "chore",
33+
"section": "Chores"
34+
},
35+
{
36+
"type": "docs",
37+
"section": "Documentation"
38+
},
39+
{
40+
"type": "style",
41+
"section": "Styles"
42+
},
43+
{
44+
"type": "refactor",
45+
"section": "Refactors"
46+
},
47+
{
48+
"type": "test",
49+
"section": "Tests",
50+
"hidden": true
51+
},
52+
{
53+
"type": "build",
54+
"section": "Build System"
55+
},
56+
{
57+
"type": "ci",
58+
"section": "Continuous Integration",
59+
"hidden": true
60+
}
61+
],
62+
"release-type": "python",
63+
"extra-files": [
64+
"src/llama_api_client/_version.py"
65+
]
66+
}

0 commit comments

Comments
 (0)