Skip to content
This repository has been archived by the owner on Sep 5, 2023. It is now read-only.

Commit

Permalink
chore: use gapic-generator-python 0.52.0 (#110)
Browse files Browse the repository at this point in the history
- [ ] Regenerate this pull request now.

fix: improper types in pagers generation
PiperOrigin-RevId: 399773015

Source-Link: googleapis/googleapis@410c184

Source-Link: googleapis/googleapis-gen@290e883
Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiMjkwZTg4MzU0NWUzYWM5ZmYyYmQwMGNkMGRhY2IyOGYxYjhjYTk0NSJ9
  • Loading branch information
gcf-owl-bot[bot] committed Sep 30, 2021
1 parent fbe6b59 commit cc27745
Show file tree
Hide file tree
Showing 3 changed files with 54 additions and 54 deletions.
Expand Up @@ -15,13 +15,13 @@
#
from typing import (
Any,
AsyncIterable,
AsyncIterator,
Awaitable,
Callable,
Iterable,
Sequence,
Tuple,
Optional,
Iterator,
)

from google.cloud.websecurityscanner_v1.types import crawled_url
Expand Down Expand Up @@ -78,14 +78,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListScanConfigsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListScanConfigsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[scan_config.ScanConfig]:
def __iter__(self) -> Iterator[scan_config.ScanConfig]:
for page in self.pages:
yield from page.scan_configs

Expand Down Expand Up @@ -142,14 +142,14 @@ def __getattr__(self, name: str) -> Any:
@property
async def pages(
self,
) -> AsyncIterable[web_security_scanner.ListScanConfigsResponse]:
) -> AsyncIterator[web_security_scanner.ListScanConfigsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[scan_config.ScanConfig]:
def __aiter__(self) -> AsyncIterator[scan_config.ScanConfig]:
async def async_generator():
async for page in self.pages:
for response in page.scan_configs:
Expand Down Expand Up @@ -208,14 +208,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListScanRunsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListScanRunsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[scan_run.ScanRun]:
def __iter__(self) -> Iterator[scan_run.ScanRun]:
for page in self.pages:
yield from page.scan_runs

Expand Down Expand Up @@ -270,14 +270,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
async def pages(self) -> AsyncIterable[web_security_scanner.ListScanRunsResponse]:
async def pages(self) -> AsyncIterator[web_security_scanner.ListScanRunsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[scan_run.ScanRun]:
def __aiter__(self) -> AsyncIterator[scan_run.ScanRun]:
async def async_generator():
async for page in self.pages:
for response in page.scan_runs:
Expand Down Expand Up @@ -336,14 +336,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListCrawledUrlsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListCrawledUrlsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[crawled_url.CrawledUrl]:
def __iter__(self) -> Iterator[crawled_url.CrawledUrl]:
for page in self.pages:
yield from page.crawled_urls

Expand Down Expand Up @@ -400,14 +400,14 @@ def __getattr__(self, name: str) -> Any:
@property
async def pages(
self,
) -> AsyncIterable[web_security_scanner.ListCrawledUrlsResponse]:
) -> AsyncIterator[web_security_scanner.ListCrawledUrlsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[crawled_url.CrawledUrl]:
def __aiter__(self) -> AsyncIterator[crawled_url.CrawledUrl]:
async def async_generator():
async for page in self.pages:
for response in page.crawled_urls:
Expand Down Expand Up @@ -466,14 +466,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListFindingsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListFindingsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[finding.Finding]:
def __iter__(self) -> Iterator[finding.Finding]:
for page in self.pages:
yield from page.findings

Expand Down Expand Up @@ -528,14 +528,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
async def pages(self) -> AsyncIterable[web_security_scanner.ListFindingsResponse]:
async def pages(self) -> AsyncIterator[web_security_scanner.ListFindingsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[finding.Finding]:
def __aiter__(self) -> AsyncIterator[finding.Finding]:
async def async_generator():
async for page in self.pages:
for response in page.findings:
Expand Down
Expand Up @@ -15,13 +15,13 @@
#
from typing import (
Any,
AsyncIterable,
AsyncIterator,
Awaitable,
Callable,
Iterable,
Sequence,
Tuple,
Optional,
Iterator,
)

from google.cloud.websecurityscanner_v1alpha.types import crawled_url
Expand Down Expand Up @@ -78,14 +78,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListScanConfigsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListScanConfigsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[scan_config.ScanConfig]:
def __iter__(self) -> Iterator[scan_config.ScanConfig]:
for page in self.pages:
yield from page.scan_configs

Expand Down Expand Up @@ -142,14 +142,14 @@ def __getattr__(self, name: str) -> Any:
@property
async def pages(
self,
) -> AsyncIterable[web_security_scanner.ListScanConfigsResponse]:
) -> AsyncIterator[web_security_scanner.ListScanConfigsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[scan_config.ScanConfig]:
def __aiter__(self) -> AsyncIterator[scan_config.ScanConfig]:
async def async_generator():
async for page in self.pages:
for response in page.scan_configs:
Expand Down Expand Up @@ -208,14 +208,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListScanRunsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListScanRunsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[scan_run.ScanRun]:
def __iter__(self) -> Iterator[scan_run.ScanRun]:
for page in self.pages:
yield from page.scan_runs

Expand Down Expand Up @@ -270,14 +270,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
async def pages(self) -> AsyncIterable[web_security_scanner.ListScanRunsResponse]:
async def pages(self) -> AsyncIterator[web_security_scanner.ListScanRunsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[scan_run.ScanRun]:
def __aiter__(self) -> AsyncIterator[scan_run.ScanRun]:
async def async_generator():
async for page in self.pages:
for response in page.scan_runs:
Expand Down Expand Up @@ -336,14 +336,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListCrawledUrlsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListCrawledUrlsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[crawled_url.CrawledUrl]:
def __iter__(self) -> Iterator[crawled_url.CrawledUrl]:
for page in self.pages:
yield from page.crawled_urls

Expand Down Expand Up @@ -400,14 +400,14 @@ def __getattr__(self, name: str) -> Any:
@property
async def pages(
self,
) -> AsyncIterable[web_security_scanner.ListCrawledUrlsResponse]:
) -> AsyncIterator[web_security_scanner.ListCrawledUrlsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[crawled_url.CrawledUrl]:
def __aiter__(self) -> AsyncIterator[crawled_url.CrawledUrl]:
async def async_generator():
async for page in self.pages:
for response in page.crawled_urls:
Expand Down Expand Up @@ -466,14 +466,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
def pages(self) -> Iterable[web_security_scanner.ListFindingsResponse]:
def pages(self) -> Iterator[web_security_scanner.ListFindingsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response

def __iter__(self) -> Iterable[finding.Finding]:
def __iter__(self) -> Iterator[finding.Finding]:
for page in self.pages:
yield from page.findings

Expand Down Expand Up @@ -528,14 +528,14 @@ def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)

@property
async def pages(self) -> AsyncIterable[web_security_scanner.ListFindingsResponse]:
async def pages(self) -> AsyncIterator[web_security_scanner.ListFindingsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response

def __aiter__(self) -> AsyncIterable[finding.Finding]:
def __aiter__(self) -> AsyncIterator[finding.Finding]:
async def async_generator():
async for page in self.pages:
for response in page.findings:
Expand Down

0 comments on commit cc27745

Please sign in to comment.