Unit Test Results.

Designed for use with JUnit and Ant.

Class dtest-novnode_jdk11_python3.8_cythonno_x86_64_32_64

NameTestsErrorsFailuresSkippedTime(s)Time StampHost
8_cythonno_x86_64_32_64180261890.4392024-11-27T12:48:41.2573669b6c40f67f0f

Tests

NameStatusTypeTime(s)
test_drop_ks_authSuccess57.598
test_permission_cache_background_reload_handle_unavailableSuccess116.353
test_defaultsSuccess57.122
test_snapshotSkippedawaiting CASSANDRA-10699

/home/cassandra/cassandra-dtest/concurrent_schema_changes_test.py:498: awaiting CASSANDRA-10699
0.002
test_tracing_simpleSuccess122.010
test_14766Skipped5.1 > 4

/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 4
1.063
test_cleanup_when_no_replica_without_indexSuccess99.198
test_single_partition_deletionsSuccess99.016
test_update_on_static_column_with_not_empty_partitionSuccess161.353
test_failed_snitch_update_property_file_snitchSkipped5.1 > 5.0.x

/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 5.0.x
1.477
test_dont_archive_commitlogFailureAssertionError: It's been over 180s and we haven't written a new commitlog segment. Something is wrong.

self = <snapshot_test.TestArchiveCommitlog object at 0x7faad0239af0>

def test_dont_archive_commitlog(self):
"""
Run the archive commitlog test, but forget to add the restore commands
"""
> self.run_archive_commitlog(restore_point_in_time=False, restore_archived_commitlog=False)

snapshot_test.py:255:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
snapshot_test.py:303: in run_archive_commitlog
advance_to_next_cl_segment(
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _

session = <cassandra.cluster.Session object at 0x7faad0014190>
commitlog_dir = '/home/cassandra/cassandra/build/run-python-dtest.T3wZLX/dtest-tael4ce2/test/node1/commitlogs'
keyspace_name = 'ks', table_name = 'junk_table', timeout = 180

def advance_to_next_cl_segment(session, commitlog_dir,
keyspace_name='ks', table_name='junk_table',
timeout=180):
"""
This is a hack to work around problems like CASSANDRA-11811.

The problem happens in commitlog-replaying tests, like the snapshot and CDC
tests. If we replay the first commitlog that's created, we wind up
replaying some mutations that initialize system tables, so this function
advances the node to the next CL by filling up the first one.
"""
session.execute(
'CREATE TABLE {ks}.{tab} ('
'a uuid PRIMARY KEY, b uuid, c uuid, d uuid, '
'e uuid, f uuid, g uuid, h uuid'
')'.format(ks=keyspace_name, tab=table_name)
)
prepared_insert = session.prepare(
'INSERT INTO {ks}.{tab} '
'(a, b, c, d, e, f, g, h) '
'VALUES ('
'uuid(), uuid(), uuid(), uuid(), '
'uuid(), uuid(), uuid(), uuid()'
')'.format(ks=keyspace_name, tab=table_name)
)

# record segments that we want to advance past
initial_cl_files = _files_in(commitlog_dir)

start = time.time()
stop_time = start + timeout
rate_limited_debug_logger = get_rate_limited_function(logger.debug, 5)
logger.debug('attempting to write until we start writing to new CL segments: {}'.format(initial_cl_files))

while _files_in(commitlog_dir) <= initial_cl_files:
elapsed = time.time() - start
rate_limited_debug_logger(' commitlog-advancing load step has lasted {s:.2f}s'.format(s=elapsed))
> assert (
time.time() <= stop_time), ("It's been over {s}s and we haven't written a new " +
"commitlog segment. Something is wrong.").format(s=timeout)
E AssertionError: It's been over 180s and we haven't written a new commitlog segment. Something is wrong.

tools/hacks.py:59: AssertionError
241.782
test_super_subcolumn_limitSkipped5.1 > 4

/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 4
1.840
test_insert_ttlSkipped5.1 > 4

/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 4
1.563
test_speculative_writeSuccess152.330
test_wide_rowsSuccess172.834
test_create_keyspaceSuccess58.566
test_round_trip_with_rate_fileSuccess3.794
test_round_trip_with_rate_fileFailurecassandra.OperationTimedOut: errors={'127.0.0.1:9042': 'Client request timeout. See Session.execute[_async](timeout)'}, last_host=127.0.0.1:9042

self = <cqlsh_tests.test_cqlsh_copy.TestCqlshCopy object at 0x7faacfeff4f0>

def test_round_trip_with_rate_file(self):
"""
Test a round trip with a large number of rows and a rate file. Make sure the rate file contains
output statistics

@jira_ticket CASSANDRA-9303
"""
num_rows = 200000
report_frequency = 0.1 # every 100 milliseconds
stress_table = 'keyspace1.standard1'
ratefile = self.get_temp_file()
tempfile = self.get_temp_file()

def check_rate_file():
lines = [line.rstrip('\n') for line in open(ratefile.name)]
logger.debug(lines)
assert lines[-1].startswith('Processed: {} rows;'.format(num_rows))

self.prepare()

logger.debug('Running stress')
self.node1.stress(['write', 'n={}'.format(num_rows), 'no-warmup', '-rate', 'threads=50'])

logger.debug('Exporting to csv file: {}'.format(tempfile.name))
self.run_cqlsh(cmds="COPY {} TO '{}' WITH RATEFILE='{}' AND REPORTFREQUENCY='{}'"
.format(stress_table, tempfile.name, ratefile.name, report_frequency))

# check all records were exported
assert num_rows == len(open(tempfile.name).readlines())

check_rate_file()

# clean-up
os.unlink(ratefile.name)
self.session.execute("TRUNCATE {}".format(stress_table))

logger.debug('Importing from csv file: {}'.format(tempfile.name))
self.run_cqlsh(cmds="COPY {} FROM '{}' WITH RATEFILE='{}' AND REPORTFREQUENCY='{}'"
.format(stress_table, tempfile.name, ratefile.name, report_frequency))

# check all records were imported
> assert [[num_rows]] == rows_to_list(self.session.execute("SELECT COUNT(*) FROM {}"
.format(stress_table)))

cqlsh_tests/test_cqlsh_copy.py:2136:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../cassandra/build/venv/src/cassandra-driver/cassandra/cluster.py:2618: in execute
return self.execute_async(query, parameters, trace, custom_payload, timeout, execution_profile, paging_state, host, execute_as).result()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _

self = <ResponseFuture: query='<SimpleStatement query="SELECT COUNT(*) FROM keyspace1.standard1", consistency=Not Set>' reque...9042': 'Client request timeout. See Session.execute[_async](timeout)'}, last_host=127.0.0.1:9042 coordinator_host=None>

def result(self):
"""
Return the final result or raise an Exception if errors were
encountered. If the final result or error has not been set
yet, this method will block until it is set, or the timeout
set for the request expires.

Timeout is specified in the Session request execution functions.
If the timeout is exceeded, an :exc:`cassandra.OperationTimedOut` will be raised.
This is a client-side timeout. For more information
about server-side coordinator timeouts, see :class:`.policies.RetryPolicy`.

Example usage::

>>> future = session.execute_async("SELECT * FROM mycf")
>>> # do other stuff...

>>> try:
... rows = future.result()
... for row in rows:
... ... # process results
... except Exception:
... log.exception("Operation failed:")

"""
self._event.wait()
if self._final_result is not _NOT_SET:
return ResultSet(self, self._final_result)
else:
> raise self._final_exception
E cassandra.OperationTimedOut: errors={'127.0.0.1:9042': 'Client request timeout. See Session.execute[_async](timeout)'}, last_host=127.0.0.1:9042

../cassandra/build/venv/src/cassandra-driver/cassandra/cluster.py:4894: OperationTimedOut
265.524
test_no_anticompaction_of_already_repairedSkipped5.1 > 4

/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 4
1.572
Properties »