diff --git a/test/streaming-sim/test_simresults.py b/test/streaming-sim/test_simresults.py index 66fb79a3442245ed5f845e385ed42a449359c80c..9c87987cc9b2aa70ae8d555b0f65da72a9e97a08 100644 --- a/test/streaming-sim/test_simresults.py +++ b/test/streaming-sim/test_simresults.py @@ -1,7 +1,6 @@ #!/usr/bin/python3 -from urllib.parse import urlencode -from urllib.request import Request, urlopen +from influxdb import InfluxDBClient import pytest from StreamingSim import run_simulation_fixture @@ -12,13 +11,18 @@ class TestSimulation(object): """ @pytest.mark.parametrize("query, expected_result", [ - ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"cpu_usage\"", "{\"results\":[{\"statement_id\":0,\"series\":[{\"name\":\"cpu_usage\",\"columns\":[\"time\",\"count_cpu_active_time\",\"count_cpu_idle_time\",\"count_cpu_usage\"],\"values\":[[\"1970-01-01T00:00:00Z\",7200,7200,7200]]}]}]}"), - ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"ipendpoint_route\"", "{\"results\":[{\"statement_id\":0,\"series\":[{\"name\":\"ipendpoint_route\",\"columns\":[\"time\",\"count_http_requests_fqdn_m\",\"count_network_fqdn_latency\"],\"values\":[[\"1970-01-01T00:00:00Z\",7200,7200]]}]}]}"), - ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"mpegdash_service\"", "{\"results\":[{\"statement_id\":0,\"series\":[{\"name\":\"mpegdash_service\",\"columns\":[\"time\",\"count_avg_response_time\",\"count_peak_response_time\",\"count_requests\"],\"values\":[[\"1970-01-01T00:00:00Z\",7200,7200,7200]]}]}]}"), - ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"net_port_io\"", "{\"results\":[{\"statement_id\":0,\"series\":[{\"name\":\"net_port_io\",\"columns\":[\"time\",\"count_RX_BYTES_PORT_M\",\"count_TX_BYTES_PORT_M\"],\"values\":[[\"1970-01-01T00:00:00Z\",7200,7200]]}]}]}"), - ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"vm_res_alloc\"", "{\"results\":[{\"statement_id\":0,\"series\":[{\"name\":\"vm_res_alloc\",\"columns\":[\"time\",\"count_cpu\",\"count_memory\",\"count_storage\"],\"values\":[[\"1970-01-01T00:00:00Z\",12,12,12]]}]}]}") + ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"cpu_usage\"", + {"time": "1970-01-01T00:00:00Z", "count_cpu_active_time": 7200, "count_cpu_idle_time": 7200, "count_cpu_usage": 7200}), + ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"ipendpoint_route\"", + {"time": "1970-01-01T00:00:00Z", "count_http_requests_fqdn_m": 7200, "count_network_fqdn_latency": 7200}), + ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"mpegdash_service\"", + {"time": "1970-01-01T00:00:00Z", "count_avg_response_time": 7200, "count_peak_response_time": 7200, "count_requests": 7200}), + ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"net_port_io\"", + {"time": "1970-01-01T00:00:00Z", "count_RX_BYTES_PORT_M": 7200, "count_TX_BYTES_PORT_M": 7200}), + ("SELECT count(*) FROM \"CLMCMetrics\".\"autogen\".\"vm_res_alloc\"", + {"time": "1970-01-01T00:00:00Z", "count_cpu": 12, "count_memory": 12, "count_storage": 12}) ]) - def test_simulation(self, query, expected_result, run_simulation_fixture): + def test_simulation(self, query, expected_result, get_db_client, run_simulation_fixture): """ This is the entry point of the test. This method will be found and executed when the module is ran using pytest @@ -27,27 +31,23 @@ class TestSimulation(object): :param run_simulation_fixture: the imported fixture to use to generate the testing data - the return value of the fixture is not needed in this case """ - print("\n") # prints a blank line for formatting purposes - # pytest automatically goes through all queries under test, declared in the parameters decorator - actual_result = self.send_query("http://localhost:8086", query) - assert expected_result == actual_result, "Simulation test failure" + print("\n") # prints a blank line for formatting purposes - print("Successfully passed test for the following query: {0}".format(query)) + # the raise_errors=False argument is given so that we could actually test that the DB didn't return any errors instead of raising an exception + query_result = get_db_client.query(query, raise_errors=False) - @staticmethod - def send_query(url, query): - """ - An auxiliary static method to send a query to a url and retrieve the result + # test the error attribute of the result is None, that is no error is returned from executing the DB query + assert query_result.error is None, "An error was encountered while executing query {0}.".format(query) - :param url: the target url to which the query is sent to - a string containing a valid URL address - :param query: the query to be executed on the given URL - :return: the result of the executed query - """ + # get the dictionary of result points; the next() function just gets the first element of the query results iterator (we only expect one item in the iterator) + actual_result = next(query_result.get_points()) - query = urlencode({"q": query}).encode("ascii") - request = Request("{0}/query".format(url), query) - result = urlopen(request) + assert expected_result == actual_result, "Simulation test failure" + + print("Successfully passed test for the following query: {0}".format(query)) - return result.read().decode("utf-8").strip() + @pytest.fixture(params=[{'database': 'CLMCMetrics'}], scope='class') + def get_db_client(self, request): + return InfluxDBClient(host='localhost', port=8086, database=request.param['database'], timeout=10)