forked from GoogleCloudPlatform/python-docs-samples
-
Notifications
You must be signed in to change notification settings - Fork 0
/
triggers_test.py
102 lines (80 loc) · 3.07 KB
/
triggers_test.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the 'License');
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an 'AS IS' BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
import google.api_core.exceptions
import google.cloud.storage
import pytest
import triggers
GCLOUD_PROJECT = os.getenv("GCLOUD_PROJECT")
TEST_BUCKET_NAME = GCLOUD_PROJECT + "-dlp-python-client-test"
RESOURCE_DIRECTORY = os.path.join(os.path.dirname(__file__), "resources")
RESOURCE_FILE_NAMES = ["test.txt", "test.png", "harmless.txt", "accounts.txt"]
TEST_TRIGGER_ID = "test-trigger"
@pytest.fixture(scope="module")
def bucket():
# Creates a GCS bucket, uploads files required for the test, and tears down
# the entire bucket afterwards.
client = google.cloud.storage.Client()
try:
bucket = client.get_bucket(TEST_BUCKET_NAME)
except google.cloud.exceptions.NotFound:
bucket = client.create_bucket(TEST_BUCKET_NAME)
# Upoad the blobs and keep track of them in a list.
blobs = []
for name in RESOURCE_FILE_NAMES:
path = os.path.join(RESOURCE_DIRECTORY, name)
blob = bucket.blob(name)
blob.upload_from_filename(path)
blobs.append(blob)
# Yield the object to the test; lines after this execute as a teardown.
yield bucket
# Delete the files.
for blob in blobs:
try:
blob.delete()
except google.cloud.exceptions.NotFound:
print("Issue during teardown, missing blob")
# Attempt to delete the bucket; this will only work if it is empty.
bucket.delete()
def test_create_list_and_delete_trigger(bucket, capsys):
try:
triggers.create_trigger(
GCLOUD_PROJECT,
bucket.name,
7,
["FIRST_NAME", "EMAIL_ADDRESS", "PHONE_NUMBER"],
trigger_id=TEST_TRIGGER_ID,
)
except google.api_core.exceptions.InvalidArgument:
# Trigger already exists, perhaps due to a previous interrupted test.
triggers.delete_trigger(GCLOUD_PROJECT, TEST_TRIGGER_ID)
out, _ = capsys.readouterr()
assert TEST_TRIGGER_ID in out
# Try again and move on.
triggers.create_trigger(
GCLOUD_PROJECT,
bucket.name,
7,
["FIRST_NAME", "EMAIL_ADDRESS", "PHONE_NUMBER"],
trigger_id=TEST_TRIGGER_ID,
auto_populate_timespan=True,
)
out, _ = capsys.readouterr()
assert TEST_TRIGGER_ID in out
triggers.list_triggers(GCLOUD_PROJECT)
out, _ = capsys.readouterr()
assert TEST_TRIGGER_ID in out
triggers.delete_trigger(GCLOUD_PROJECT, TEST_TRIGGER_ID)
out, _ = capsys.readouterr()
assert TEST_TRIGGER_ID in out