229 lines
8.7 KiB
Python
Executable File
229 lines
8.7 KiB
Python
Executable File
import logging
|
|
import time
|
|
import typing
|
|
from datetime import datetime
|
|
|
|
import allure
|
|
import cv2
|
|
import numpy as np
|
|
import pytesseract
|
|
from PIL import ImageGrab
|
|
|
|
import configs
|
|
import constants
|
|
import driver
|
|
from configs.system import IS_LIN
|
|
from scripts.tools.ocv import Ocv
|
|
from scripts.utils.system_path import SystemPath
|
|
|
|
_logger = logging.getLogger(__name__)
|
|
|
|
|
|
class Image:
|
|
|
|
def __init__(self, object_name: dict):
|
|
self.object_name = object_name
|
|
self._view = None
|
|
|
|
@property
|
|
@allure.step('Get image view')
|
|
def view(self) -> np.ndarray:
|
|
return self._view
|
|
|
|
@property
|
|
@allure.step('Get image height')
|
|
def height(self) -> int:
|
|
return self.view.shape[0]
|
|
|
|
@property
|
|
@allure.step('Get image width')
|
|
def width(self) -> int:
|
|
return self.view.shape[1]
|
|
|
|
@property
|
|
@allure.step('Get image is grayscale')
|
|
def is_grayscale(self) -> bool:
|
|
return self.view.ndim == 2
|
|
|
|
@allure.step('Set image in grayscale')
|
|
def set_grayscale(self) -> 'Image':
|
|
if not self.is_grayscale:
|
|
self._view = cv2.cvtColor(self.view, cv2.COLOR_BGR2GRAY)
|
|
return self
|
|
|
|
@allure.step('Grab image view from object')
|
|
def update_view(self):
|
|
_logger.debug(f'Image view was grab from: {self.object_name}')
|
|
rect = driver.object.globalBounds(driver.waitForObject(self.object_name))
|
|
img = ImageGrab.grab(
|
|
bbox=(rect.x, rect.y, rect.x + rect.width, rect.y + rect.height),
|
|
xdisplay=":0" if IS_LIN else None
|
|
)
|
|
self._view = cv2.cvtColor(np.array(img), cv2.COLOR_BGR2RGB)
|
|
|
|
@allure.step('Save image')
|
|
def save(self, path: SystemPath, force: bool = False):
|
|
path.parent.mkdir(parents=True, exist_ok=True)
|
|
if path.exists() and not force:
|
|
raise FileExistsError(path)
|
|
if self.view is None:
|
|
self.update_view()
|
|
cv2.imwrite(str(path), self.view)
|
|
|
|
@allure.step('Compare images')
|
|
def compare(
|
|
self, expected: np.ndarray, threshold: float = 0.99) -> bool:
|
|
if self.view is None:
|
|
self.update_view()
|
|
correlation = Ocv.compare_images(self.view, expected)
|
|
result = correlation >= threshold
|
|
_logger.info(f'Images equals on: {abs(round(correlation, 4) * 100)}%')
|
|
|
|
if result:
|
|
_logger.info(f'Screenshot comparison passed')
|
|
else:
|
|
configs.testpath.TEST_ARTIFACTS.mkdir(parents=True, exist_ok=True)
|
|
diff = Ocv.draw_contours(self.view, expected)
|
|
|
|
actual_fp = configs.testpath.TEST_ARTIFACTS / f'actual_image.png'
|
|
expected_fp = configs.testpath.TEST_ARTIFACTS / f'expected_image.png'
|
|
diff_fp = configs.testpath.TEST_ARTIFACTS / f'diff_image.png'
|
|
|
|
self.save(actual_fp, force=True)
|
|
cv2.imwrite(str(expected_fp), expected)
|
|
cv2.imwrite(str(diff_fp), diff)
|
|
|
|
allure.attach(name='actual', body=actual_fp.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
allure.attach(name='expected', body=expected_fp.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
allure.attach(name='diff', body=diff_fp.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
|
|
_logger.info(
|
|
f"Screenshot comparison failed.\n"
|
|
f"Actual, Diff and Expected screenshots are saved:\n"
|
|
f"{configs.testpath.TEST_ARTIFACTS.relative_to(configs.testpath.ROOT)}.")
|
|
return result
|
|
|
|
@allure.step('Crop image')
|
|
def crop(self, rect: driver.UiTypes.ScreenRectangle):
|
|
assert rect.x + rect.width < self.width
|
|
assert rect.y + rect.height < self.height
|
|
self._view = self.view[rect.y: (rect.y + rect.height), rect.x: (rect.x + rect.width)]
|
|
|
|
@allure.step('Parse text on image')
|
|
def to_string(self, custom_config: str):
|
|
text: str = pytesseract.image_to_string(self.view, config=custom_config)
|
|
_logger.debug(f'Text on image: {text}')
|
|
return text
|
|
|
|
@allure.step('Verify: Image contains text: {1}')
|
|
def has_text(self, text: str, criteria: str, crop: driver.UiTypes.ScreenRectangle = None) -> bool:
|
|
self.update_view()
|
|
if crop:
|
|
self.crop(crop)
|
|
|
|
# Search text on image converted in gray color
|
|
self.set_grayscale()
|
|
fp_gray = configs.testpath.TEST_ARTIFACTS / f'search_region_in_gray_color.png'
|
|
self.save(fp_gray, force=True)
|
|
if text.lower() in self.to_string(criteria).lower():
|
|
allure.attach(name='search_region', body=fp_gray.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
return True
|
|
|
|
# Search text on image with inverted color
|
|
self._view = cv2.bitwise_not(self.view)
|
|
fp_invert = configs.testpath.TEST_ARTIFACTS / f'search_region_in_inverted_color.png'
|
|
self.save(fp_invert, force=True)
|
|
if text.lower() in self.to_string(criteria).lower():
|
|
allure.attach(name='search_region', body=fp_invert.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
return True
|
|
return False
|
|
|
|
@allure.step('Search color on image')
|
|
def has_color(self, color: constants.Color, denoise: int = 10, crop: driver.UiTypes.ScreenRectangle = None) -> bool:
|
|
self.update_view()
|
|
if crop:
|
|
self.crop(crop)
|
|
|
|
initial_view = configs.testpath.TEST_ARTIFACTS / f'{color.name}.png'
|
|
self.save(initial_view)
|
|
allure.attach(name='search_region', body=initial_view.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
|
|
contours = self._get_color_contours(color, denoise, apply=True)
|
|
|
|
mask_view = configs.testpath.TEST_ARTIFACTS / f'{color.name}_mask.png'
|
|
self.save(mask_view)
|
|
allure.attach(name='contours', body=mask_view.read_bytes(), attachment_type=allure.attachment_type.PNG)
|
|
|
|
self._view = None
|
|
return len(contours) >= 1
|
|
|
|
@allure.step('Apply contours with found color on image')
|
|
def _get_color_contours(
|
|
self,
|
|
color: constants.Color,
|
|
denoise: int = 10,
|
|
apply: bool = False
|
|
) -> typing.List[driver.UiTypes.ScreenRectangle]:
|
|
if not self.is_grayscale:
|
|
view = cv2.cvtColor(self.view, cv2.COLOR_BGR2HSV)
|
|
else:
|
|
view = self.view
|
|
boundaries = constants.boundaries[color]
|
|
|
|
if color == constants.Color.RED:
|
|
mask = None
|
|
for bond in boundaries:
|
|
lower_range = np.array(bond[0])
|
|
upper_range = np.array(bond[1])
|
|
_mask = cv2.inRange(view, lower_range, upper_range)
|
|
mask = _mask if mask is None else mask + _mask
|
|
else:
|
|
lower_range = np.array(boundaries[0])
|
|
upper_range = np.array(boundaries[1])
|
|
mask = cv2.inRange(view, lower_range, upper_range)
|
|
|
|
contours = []
|
|
_contours = cv2.findContours(mask, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)
|
|
_contours = _contours[0] if len(_contours) == 2 else _contours[1]
|
|
for _contour in _contours:
|
|
x, y, w, h = cv2.boundingRect(_contour)
|
|
# To remove small artifacts, less than denoise variable value
|
|
if w * h < denoise:
|
|
continue
|
|
contours.append(driver.UiTypes.ScreenRectangle(x, y, w, h))
|
|
|
|
if apply:
|
|
self._view = cv2.bitwise_and(self.view, self.view, mask=mask)
|
|
for contour in contours:
|
|
cv2.rectangle(
|
|
self.view,
|
|
(contour.x, contour.y),
|
|
(contour.x + contour.width, contour.y + contour.height),
|
|
(36, 255, 12), 2)
|
|
|
|
return contours
|
|
|
|
|
|
@allure.step('Compare images')
|
|
def compare(actual: Image,
|
|
expected: typing.Union[str, SystemPath, Image],
|
|
threshold: float = 0.99,
|
|
timout_sec: int = 1
|
|
):
|
|
if isinstance(expected, str):
|
|
expected_fp = configs.testpath.TEST_VP / configs.system.OS_ID / expected
|
|
if not expected_fp.exists():
|
|
expected_fp = configs.testpath.TEST_VP / expected
|
|
expected = expected_fp
|
|
if isinstance(expected, SystemPath):
|
|
assert expected.exists(), f'File: {expected} not found'
|
|
expected = cv2.imread(str(expected))
|
|
else:
|
|
expected = expected.view
|
|
start = datetime.now()
|
|
while not actual.compare(expected, threshold):
|
|
time.sleep(1)
|
|
assert (datetime.now() - start).seconds < timout_sec, 'Comparison failed'
|
|
_logger.info(f'Screenshot comparison passed')
|
|
|