summaryrefslogtreecommitdiffstats
path: root/lldb/packages/Python/lldbsuite/test/tools/lldb-server
diff options
context:
space:
mode:
Diffstat (limited to 'lldb/packages/Python/lldbsuite/test/tools/lldb-server')
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/.clang-format1
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/Makefile8
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestAppleSimulatorOSType.py123
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAttach.py67
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAuxvSupport.py220
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExitCode.py127
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExpeditedRegisters.py162
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteHostInfo.py131
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteKill.py59
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteModuleInfo.py44
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteProcessInfo.py211
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteRegisterState.py128
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteSingleStep.py41
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteThreadsInStopReply.py303
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_qThreadStopInfo.py182
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_vCont.py159
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestLldbGdbServer.py1570
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubReverseConnect.py98
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubSetSID.py86
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/gdbremote_testcase.py1643
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/Makefile8
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteAbort.py46
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteSegFault.py46
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/main.cpp31
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/lldbgdbserverutils.py945
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/main.cpp370
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/Makefile5
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/TestPlatformProcessConnect.py96
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/main.cpp6
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/Makefile5
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/TestGdbRemoteGPacket.py153
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/main.cpp54
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/Makefile5
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/TestGdbRemote_QPassSignals.py115
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/main.cpp36
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/socket_packet_pump.py198
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/test/test_lldbgdbserverutils.py65
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/Makefile6
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/TestGdbRemoteThreadName.py41
-rw-r--r--lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/main.cpp22
40 files changed, 0 insertions, 7616 deletions
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/.clang-format b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/.clang-format
deleted file mode 100644
index 9b3aa8b7213..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/.clang-format
+++ /dev/null
@@ -1 +0,0 @@
-BasedOnStyle: LLVM
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/Makefile
deleted file mode 100644
index 28aba3cf546..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/Makefile
+++ /dev/null
@@ -1,8 +0,0 @@
-LEVEL = ../../make
-
-override CFLAGS_EXTRAS += -D__STDC_LIMIT_MACROS -D__STDC_FORMAT_MACROS
-ENABLE_THREADS := YES
-CXX_SOURCES := main.cpp
-MAKE_DSYM :=NO
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestAppleSimulatorOSType.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestAppleSimulatorOSType.py
deleted file mode 100644
index ab01e44db4b..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestAppleSimulatorOSType.py
+++ /dev/null
@@ -1,123 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import lldbgdbserverutils
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-import json
-
-class TestAppleSimulatorOSType(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def check_simulator_ostype(self, sdk, platform, arch='x86_64'):
- sim_devices_str = subprocess.check_output(['xcrun', 'simctl', 'list',
- '-j', 'devices']).decode("utf-8")
- sim_devices = json.loads(sim_devices_str)['devices']
- # Find an available simulator for the requested platform
- deviceUDID = None
- for simulator in sim_devices:
- if isinstance(simulator,dict):
- runtime = simulator['name']
- devices = simulator['devices']
- else:
- runtime = simulator
- devices = sim_devices[simulator]
- if not platform in runtime.lower():
- continue
- for device in devices:
- if 'availability' in device and device['availability'] != '(available)':
- continue
- if 'isAvailable' in device and device['isAvailable'] != True:
- continue
- deviceUDID = device['udid']
- break
- if deviceUDID != None:
- break
-
- # Launch the process using simctl
- self.assertIsNotNone(deviceUDID)
- exe_name = 'test_simulator_platform_{}'.format(platform)
- sdkroot = subprocess.check_output(['xcrun', '--show-sdk-path', '--sdk',
- sdk]).decode("utf-8")
- self.build(dictionary={ 'EXE': exe_name, 'SDKROOT': sdkroot.strip(),
- 'ARCH': arch })
- exe_path = self.getBuildArtifact(exe_name)
- sim_launcher = subprocess.Popen(['xcrun', 'simctl', 'spawn',
- deviceUDID, exe_path,
- 'print-pid', 'sleep:10'],
- stderr=subprocess.PIPE)
- # Get the PID from the process output
- pid = None
- while not pid:
- stderr = sim_launcher.stderr.readline().decode("utf-8")
- if stderr == '':
- continue
- m = re.match(r"PID: (.*)", stderr)
- self.assertIsNotNone(m)
- pid = int(m.group(1))
-
- # Launch debug monitor attaching to the simulated process
- self.init_debugserver_test()
- server = self.connect_to_debug_monitor(attach_pid=pid)
-
- # Setup packet sequences
- self.add_no_ack_remote_stream()
- self.add_process_info_collection_packets()
- self.test_sequence.add_log_lines(
- ["read packet: " +
- "$jGetLoadedDynamicLibrariesInfos:{\"fetch_all_solibs\" : true}]#ce",
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "dylib_info_raw"}}],
- True)
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Check that ostype is correct
- self.assertEquals(process_info['ostype'], platform)
-
- # Now for dylibs
- dylib_info_raw = context.get("dylib_info_raw")
- dylib_info = json.loads(self.decode_gdbremote_binary(dylib_info_raw))
- images = dylib_info['images']
-
- image_info = None
- for image in images:
- if image['pathname'] != exe_path:
- continue
- image_info = image
- break
-
- self.assertIsNotNone(image_info)
- self.assertEquals(image['min_version_os_name'], platform)
-
-
- @apple_simulator_test('iphone')
- @debugserver_test
- @skipIfDarwinEmbedded
- def test_simulator_ostype_ios(self):
- self.check_simulator_ostype(sdk='iphonesimulator',
- platform='ios')
-
- @apple_simulator_test('appletv')
- @debugserver_test
- @skipIfDarwinEmbedded
- def test_simulator_ostype_tvos(self):
- self.check_simulator_ostype(sdk='appletvsimulator',
- platform='tvos')
-
- @apple_simulator_test('watch')
- @debugserver_test
- @skipIfDarwinEmbedded
- def test_simulator_ostype_watchos(self):
- self.check_simulator_ostype(sdk='watchsimulator',
- platform='watchos', arch='i386')
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAttach.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAttach.py
deleted file mode 100644
index 1696b599de6..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAttach.py
+++ /dev/null
@@ -1,67 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import lldbgdbserverutils
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteAttach(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def attach_with_vAttach(self):
- # Start the inferior, start the debug monitor, nothing is attached yet.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["sleep:60"])
- self.assertIsNotNone(procs)
-
- # Make sure the target process has been launched.
- inferior = procs.get("inferior")
- self.assertIsNotNone(inferior)
- self.assertTrue(inferior.pid > 0)
- self.assertTrue(
- lldbgdbserverutils.process_is_running(
- inferior.pid, True))
-
- # Add attach packets.
- self.test_sequence.add_log_lines([
- # Do the attach.
- "read packet: $vAttach;{:x}#00".format(inferior.pid),
- # Expect a stop notification from the attach.
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2})[^#]*#[0-9a-fA-F]{2}$",
- "capture": {1: "stop_signal_hex"}},
- ], True)
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the process id matches what we expected.
- pid_text = process_info.get('pid', None)
- self.assertIsNotNone(pid_text)
- reported_pid = int(pid_text, base=16)
- self.assertEqual(reported_pid, inferior.pid)
-
- @debugserver_test
- def test_attach_with_vAttach_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach_manually()
- self.attach_with_vAttach()
-
- @llgs_test
- def test_attach_with_vAttach_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach_manually()
- self.attach_with_vAttach()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAuxvSupport.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAuxvSupport.py
deleted file mode 100644
index 3c87ee61ab6..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteAuxvSupport.py
+++ /dev/null
@@ -1,220 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteAuxvSupport(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- AUXV_SUPPORT_FEATURE_NAME = "qXfer:auxv:read"
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def has_auxv_support(self):
- inferior_args = ["message:main entered", "sleep:5"]
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
-
- # Don't do anything until we match the launched inferior main entry output.
- # Then immediately interrupt the process.
- # This prevents auxv data being asked for before it's ready and leaves
- # us in a stopped state.
- self.test_sequence.add_log_lines([
- # Start the inferior...
- "read packet: $c#63",
- # ... match output....
- {"type": "output_match", "regex": self.maybe_strict_output_regex(
- r"message:main entered\r\n")},
- ], True)
- # ... then interrupt.
- self.add_interrupt_packets()
- self.add_qSupported_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- features = self.parse_qSupported_response(context)
- return self.AUXV_SUPPORT_FEATURE_NAME in features and features[
- self.AUXV_SUPPORT_FEATURE_NAME] == "+"
-
- def get_raw_auxv_data(self):
- # Start up llgs and inferior, and check for auxv support.
- if not self.has_auxv_support():
- self.skipTest("auxv data not supported")
-
- # Grab pointer size for target. We'll assume that is equivalent to an unsigned long on the target.
- # Auxv is specified in terms of pairs of unsigned longs.
- self.reset_test_sequence()
- self.add_process_info_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- proc_info = self.parse_process_info_response(context)
- self.assertIsNotNone(proc_info)
- self.assertTrue("ptrsize" in proc_info)
- word_size = int(proc_info["ptrsize"])
-
- OFFSET = 0
- LENGTH = 0x400
-
- # Grab the auxv data.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: $qXfer:auxv:read::{:x},{:x}:#00".format(
- OFFSET,
- LENGTH),
- {
- "direction": "send",
- "regex": re.compile(
- r"^\$([^E])(.*)#[0-9a-fA-F]{2}$",
- re.MULTILINE | re.DOTALL),
- "capture": {
- 1: "response_type",
- 2: "content_raw"}}],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Ensure we end up with all auxv data in one packet.
- # FIXME don't assume it all comes back in one packet.
- self.assertEqual(context.get("response_type"), "l")
-
- # Decode binary data.
- content_raw = context.get("content_raw")
- self.assertIsNotNone(content_raw)
- return (word_size, self.decode_gdbremote_binary(content_raw))
-
- def supports_auxv(self):
- # When non-auxv platforms support llgs, skip the test on platforms
- # that don't support auxv.
- self.assertTrue(self.has_auxv_support())
-
- #
- # We skip the "supports_auxv" test on debugserver. The rest of the tests
- # appropriately skip the auxv tests if the support flag is not present
- # in the qSupported response, so the debugserver test bits are still there
- # in case debugserver code one day does have auxv support and thus those
- # tests don't get skipped.
- #
-
- @llgs_test
- def test_supports_auxv_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.supports_auxv()
-
- def auxv_data_is_correct_size(self):
- (word_size, auxv_data) = self.get_raw_auxv_data()
- self.assertIsNotNone(auxv_data)
-
- # Ensure auxv data is a multiple of 2*word_size (there should be two
- # unsigned long fields per auxv entry).
- self.assertEqual(len(auxv_data) % (2 * word_size), 0)
- # print("auxv contains {} entries".format(len(auxv_data) / (2*word_size)))
-
- @debugserver_test
- def test_auxv_data_is_correct_size_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_data_is_correct_size()
-
- @llgs_test
- def test_auxv_data_is_correct_size_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_data_is_correct_size()
-
- def auxv_keys_look_valid(self):
- (word_size, auxv_data) = self.get_raw_auxv_data()
- self.assertIsNotNone(auxv_data)
-
- # Grab endian.
- self.reset_test_sequence()
- self.add_process_info_collection_packets()
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- auxv_dict = self.build_auxv_dict(endian, word_size, auxv_data)
- self.assertIsNotNone(auxv_dict)
-
- # Verify keys look reasonable.
- for auxv_key in auxv_dict:
- self.assertTrue(auxv_key >= 1)
- self.assertTrue(auxv_key <= 1000)
- # print("auxv dict: {}".format(auxv_dict))
-
- @debugserver_test
- def test_auxv_keys_look_valid_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_keys_look_valid()
-
- @llgs_test
- def test_auxv_keys_look_valid_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_keys_look_valid()
-
- def auxv_chunked_reads_work(self):
- # Verify that multiple smaller offset,length reads of auxv data
- # return the same data as a single larger read.
-
- # Grab the auxv data with a single large read here.
- (word_size, auxv_data) = self.get_raw_auxv_data()
- self.assertIsNotNone(auxv_data)
-
- # Grab endian.
- self.reset_test_sequence()
- self.add_process_info_collection_packets()
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- auxv_dict = self.build_auxv_dict(endian, word_size, auxv_data)
- self.assertIsNotNone(auxv_dict)
-
- iterated_auxv_data = self.read_binary_data_in_chunks(
- "qXfer:auxv:read::", 2 * word_size)
- self.assertIsNotNone(iterated_auxv_data)
-
- auxv_dict_iterated = self.build_auxv_dict(
- endian, word_size, iterated_auxv_data)
- self.assertIsNotNone(auxv_dict_iterated)
-
- # Verify both types of data collection returned same content.
- self.assertEqual(auxv_dict_iterated, auxv_dict)
-
- @debugserver_test
- def test_auxv_chunked_reads_work_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_chunked_reads_work()
-
- @llgs_test
- def test_auxv_chunked_reads_work_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.auxv_chunked_reads_work()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExitCode.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExitCode.py
deleted file mode 100644
index 5ef4249bd24..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExitCode.py
+++ /dev/null
@@ -1,127 +0,0 @@
-from __future__ import print_function
-
-# lldb test suite imports
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import TestBase
-
-# gdb-remote-specific imports
-import lldbgdbserverutils
-from gdbremote_testcase import GdbRemoteTestCaseBase
-
-
-class TestGdbRemoteExitCode(GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- FAILED_LAUNCH_CODE = "E08"
-
- def get_launch_fail_reason(self):
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $qLaunchSuccess#00"],
- True)
- self.test_sequence.add_log_lines(
- [{"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "launch_result"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- return context.get("launch_result")[1:]
-
- def start_inferior(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.test_sequence.add_log_lines(
- ["read packet: %s" % lldbgdbserverutils.build_gdbremote_A_packet(
- launch_args)],
- True)
- self.test_sequence.add_log_lines(
- [{"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "A_result"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- launch_result = context.get("A_result")
- self.assertIsNotNone(launch_result)
- if launch_result == self.FAILED_LAUNCH_CODE:
- fail_reason = self.get_launch_fail_reason()
- self.fail("failed to launch inferior: " + fail_reason)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_start_inferior_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.start_inferior()
-
- @llgs_test
- def test_start_inferior_llgs(self):
- self.init_llgs_test()
- self.build()
- self.start_inferior()
-
- def inferior_exit_0(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8",
- "send packet: $W00#00"],
- True)
-
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_inferior_exit_0_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.inferior_exit_0()
-
- @llgs_test
- def test_inferior_exit_0_llgs(self):
- self.init_llgs_test()
- self.build()
- self.inferior_exit_0()
-
- def inferior_exit_42(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- RETVAL = 42
-
- # build launch args
- launch_args += ["retval:%d" % RETVAL]
-
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8",
- "send packet: $W{0:02x}#00".format(RETVAL)],
- True)
-
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_inferior_exit_42_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.inferior_exit_42()
-
- @llgs_test
- def test_inferior_exit_42_llgs(self):
- self.init_llgs_test()
- self.build()
- self.inferior_exit_42()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExpeditedRegisters.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExpeditedRegisters.py
deleted file mode 100644
index 7d8e28c745c..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteExpeditedRegisters.py
+++ /dev/null
@@ -1,162 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteExpeditedRegisters(
- gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
-
- def gather_expedited_registers(self):
- # Setup the stub and set the gdb remote command stream.
- procs = self.prep_debug_monitor_and_inferior(inferior_args=["sleep:2"])
- self.test_sequence.add_log_lines([
- # Start up the inferior.
- "read packet: $c#63",
- # Immediately tell it to stop. We want to see what it reports.
- "read packet: {}".format(chr(3)),
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {1: "stop_result",
- 2: "key_vals_text"}},
- ], True)
-
- # Run the gdb remote command stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Pull out expedited registers.
- key_vals_text = context.get("key_vals_text")
- self.assertIsNotNone(key_vals_text)
-
- expedited_registers = self.extract_registers_from_stop_notification(
- key_vals_text)
- self.assertIsNotNone(expedited_registers)
-
- return expedited_registers
-
- def stop_notification_contains_generic_register(
- self, generic_register_name):
- # Generate a stop reply, parse out expedited registers from stop
- # notification.
- expedited_registers = self.gather_expedited_registers()
- self.assertIsNotNone(expedited_registers)
- self.assertTrue(len(expedited_registers) > 0)
-
- # Gather target register infos.
- reg_infos = self.gather_register_infos()
-
- # Find the generic register.
- reg_info = self.find_generic_register_with_name(
- reg_infos, generic_register_name)
- self.assertIsNotNone(reg_info)
-
- # Ensure the expedited registers contained it.
- self.assertTrue(reg_info["lldb_register_index"] in expedited_registers)
- # print("{} reg_info:{}".format(generic_register_name, reg_info))
-
- def stop_notification_contains_any_registers(self):
- # Generate a stop reply, parse out expedited registers from stop
- # notification.
- expedited_registers = self.gather_expedited_registers()
- # Verify we have at least one expedited register.
- self.assertTrue(len(expedited_registers) > 0)
-
- @debugserver_test
- def test_stop_notification_contains_any_registers_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_any_registers()
-
- @llgs_test
- def test_stop_notification_contains_any_registers_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_any_registers()
-
- def stop_notification_contains_no_duplicate_registers(self):
- # Generate a stop reply, parse out expedited registers from stop
- # notification.
- expedited_registers = self.gather_expedited_registers()
- # Verify no expedited register was specified multiple times.
- for (reg_num, value) in list(expedited_registers.items()):
- if (isinstance(value, list)) and (len(value) > 0):
- self.fail(
- "expedited register number {} specified more than once ({} times)".format(
- reg_num, len(value)))
-
- @debugserver_test
- def test_stop_notification_contains_no_duplicate_registers_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_no_duplicate_registers()
-
- @llgs_test
- def test_stop_notification_contains_no_duplicate_registers_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_no_duplicate_registers()
-
- def stop_notification_contains_pc_register(self):
- self.stop_notification_contains_generic_register("pc")
-
- @debugserver_test
- def test_stop_notification_contains_pc_register_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_pc_register()
-
- @llgs_test
- def test_stop_notification_contains_pc_register_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_pc_register()
-
- # powerpc64 has no FP register
- @skipIf(triple='^powerpc64')
- def stop_notification_contains_fp_register(self):
- self.stop_notification_contains_generic_register("fp")
-
- @debugserver_test
- def test_stop_notification_contains_fp_register_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_fp_register()
-
- @llgs_test
- def test_stop_notification_contains_fp_register_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_fp_register()
-
- def stop_notification_contains_sp_register(self):
- self.stop_notification_contains_generic_register("sp")
-
- @debugserver_test
- def test_stop_notification_contains_sp_register_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_sp_register()
-
- @llgs_test
- def test_stop_notification_contains_sp_register_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_notification_contains_sp_register()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteHostInfo.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteHostInfo.py
deleted file mode 100644
index 464cdce5e9e..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteHostInfo.py
+++ /dev/null
@@ -1,131 +0,0 @@
-from __future__ import print_function
-
-# lldb test suite imports
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import TestBase
-
-# gdb-remote-specific imports
-import lldbgdbserverutils
-from gdbremote_testcase import GdbRemoteTestCaseBase
-
-
-class TestGdbRemoteHostInfo(GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- KNOWN_HOST_INFO_KEYS = set([
- "arch",
- "cputype",
- "cpusubtype",
- "distribution_id",
- "endian",
- "hostname",
- "ostype",
- "os_build",
- "os_kernel",
- "os_version",
- "ptrsize",
- "triple",
- "vendor",
- "watchpoint_exceptions_received",
- "default_packet_timeout",
- ])
-
- DARWIN_REQUIRED_HOST_INFO_KEYS = set([
- "cputype",
- "cpusubtype",
- "endian",
- "ostype",
- "ptrsize",
- "vendor",
- "watchpoint_exceptions_received"
- ])
-
- def add_host_info_collection_packets(self):
- self.test_sequence.add_log_lines(
- ["read packet: $qHostInfo#9b",
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "host_info_raw"}}],
- True)
-
- def parse_host_info_response(self, context):
- # Ensure we have a host info response.
- self.assertIsNotNone(context)
- host_info_raw = context.get("host_info_raw")
- self.assertIsNotNone(host_info_raw)
-
- # Pull out key:value; pairs.
- host_info_dict = {match.group(1): match.group(2)
- for match in re.finditer(r"([^:]+):([^;]+);",
- host_info_raw)}
-
- import pprint
- print("\nqHostInfo response:")
- pprint.pprint(host_info_dict)
-
- # Validate keys are known.
- for (key, val) in list(host_info_dict.items()):
- self.assertTrue(key in self.KNOWN_HOST_INFO_KEYS,
- "unknown qHostInfo key: " + key)
- self.assertIsNotNone(val)
-
- # Return the key:val pairs.
- return host_info_dict
-
- def get_qHostInfo_response(self):
- # Launch the debug monitor stub, attaching to the inferior.
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
- self.add_no_ack_remote_stream()
-
- # Request qHostInfo and get response
- self.add_host_info_collection_packets()
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Parse qHostInfo response.
- host_info = self.parse_host_info_response(context)
- self.assertIsNotNone(host_info)
- self.assertGreater(len(host_info), 0, "qHostInfo should have returned "
- "at least one key:val pair.")
- return host_info
-
- def validate_darwin_minimum_host_info_keys(self, host_info_dict):
- self.assertIsNotNone(host_info_dict)
- missing_keys = [key for key in self.DARWIN_REQUIRED_HOST_INFO_KEYS
- if key not in host_info_dict]
- self.assertEquals(0, len(missing_keys),
- "qHostInfo is missing the following required "
- "keys: " + str(missing_keys))
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_qHostInfo_returns_at_least_one_key_val_pair_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.get_qHostInfo_response()
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @llgs_test
- def test_qHostInfo_returns_at_least_one_key_val_pair_llgs(self):
- self.init_llgs_test()
- self.build()
- self.get_qHostInfo_response()
-
- @skipUnlessDarwin
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_qHostInfo_contains_darwin_required_keys_debugserver(self):
- self.init_debugserver_test()
- self.build()
- host_info_dict = self.get_qHostInfo_response()
- self.validate_darwin_minimum_host_info_keys(host_info_dict)
-
- @skipUnlessDarwin
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @llgs_test
- def test_qHostInfo_contains_darwin_required_keys_llgs(self):
- self.init_llgs_test()
- self.build()
- host_info_dict = self.get_qHostInfo_response()
- self.validate_darwin_minimum_host_info_keys(host_info_dict)
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteKill.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteKill.py
deleted file mode 100644
index 6b8ed3b4c50..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteKill.py
+++ /dev/null
@@ -1,59 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import lldbgdbserverutils
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteKill(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
-
- def attach_commandline_kill_after_initial_stop(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.test_sequence.add_log_lines([
- "read packet: $k#6b",
- {"direction": "send", "regex": r"^\$X[0-9a-fA-F]+([^#]*)#[0-9A-Fa-f]{2}"},
- ], True)
-
- if self.stub_sends_two_stop_notifications_on_kill:
- # Add an expectation for a second X result for stubs that send two
- # of these.
- self.test_sequence.add_log_lines([
- {"direction": "send", "regex": r"^\$X[0-9a-fA-F]+([^#]*)#[0-9A-Fa-f]{2}"},
- ], True)
-
- self.expect_gdbremote_sequence()
-
- # Wait a moment for completed and now-detached inferior process to
- # clear.
- time.sleep(1)
-
- if not lldb.remote_platform:
- # Process should be dead now. Reap results.
- poll_result = procs["inferior"].poll()
- self.assertIsNotNone(poll_result)
-
- # Where possible, verify at the system level that the process is not
- # running.
- self.assertFalse(
- lldbgdbserverutils.process_is_running(
- procs["inferior"].pid, False))
-
- @debugserver_test
- def test_attach_commandline_kill_after_initial_stop_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_kill_after_initial_stop()
-
- @llgs_test
- def test_attach_commandline_kill_after_initial_stop_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_kill_after_initial_stop()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteModuleInfo.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteModuleInfo.py
deleted file mode 100644
index ef96b55b90d..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteModuleInfo.py
+++ /dev/null
@@ -1,44 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import lldbgdbserverutils
-from lldbsuite.support import seven
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteModuleInfo(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def module_info(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_process_info_collection_packets()
- context = self.expect_gdbremote_sequence()
- info = self.parse_process_info_response(context)
-
- self.test_sequence.add_log_lines([
- 'read packet: $jModulesInfo:[{"file":"%s","triple":"%s"}]]#00' % (
- lldbutil.append_to_process_working_directory(self, "a.out"),
- seven.unhexlify(info["triple"])),
- {"direction": "send",
- "regex": r'^\$\[{(.*)}\]\]#[0-9A-Fa-f]{2}',
- "capture": {1: "spec"}},
- ], True)
-
- context = self.expect_gdbremote_sequence()
- spec = context.get("spec")
- self.assertRegexpMatches(spec, '"file_path":".*"')
- self.assertRegexpMatches(spec, '"file_offset":\d+')
- self.assertRegexpMatches(spec, '"file_size":\d+')
- self.assertRegexpMatches(spec, '"triple":"\w*-\w*-.*"')
- self.assertRegexpMatches(spec, '"uuid":"[A-Fa-f0-9]+"')
-
- @llgs_test
- def test_module_info(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.module_info()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteProcessInfo.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteProcessInfo.py
deleted file mode 100644
index 4f7123d789a..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteProcessInfo.py
+++ /dev/null
@@ -1,211 +0,0 @@
-from __future__ import print_function
-
-
-import sys
-
-import gdbremote_testcase
-import lldbgdbserverutils
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteProcessInfo(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def qProcessInfo_returns_running_process(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the process id looks reasonable.
- pid_text = process_info.get("pid")
- self.assertIsNotNone(pid_text)
- pid = int(pid_text, base=16)
- self.assertNotEqual(0, pid)
-
- # If possible, verify that the process is running.
- self.assertTrue(lldbgdbserverutils.process_is_running(pid, True))
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qProcessInfo_returns_running_process_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.qProcessInfo_returns_running_process()
-
- @llgs_test
- def test_qProcessInfo_returns_running_process_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qProcessInfo_returns_running_process()
-
- def attach_commandline_qProcessInfo_reports_correct_pid(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.assertIsNotNone(procs)
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence(timeout_seconds=8)
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the process id matches what we expected.
- pid_text = process_info.get('pid', None)
- self.assertIsNotNone(pid_text)
- reported_pid = int(pid_text, base=16)
- self.assertEqual(reported_pid, procs["inferior"].pid)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_attach_commandline_qProcessInfo_reports_correct_pid_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_qProcessInfo_reports_correct_pid()
-
- @llgs_test
- def test_attach_commandline_qProcessInfo_reports_correct_pid_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_qProcessInfo_reports_correct_pid()
-
- def qProcessInfo_reports_valid_endian(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the process id looks reasonable.
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
- self.assertTrue(endian in ["little", "big", "pdp"])
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qProcessInfo_reports_valid_endian_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.qProcessInfo_reports_valid_endian()
-
- @llgs_test
- def test_qProcessInfo_reports_valid_endian_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qProcessInfo_reports_valid_endian()
-
- def qProcessInfo_contains_keys(self, expected_key_set):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the expected keys are present and non-None within the process
- # info.
- missing_key_set = set()
- for expected_key in expected_key_set:
- if expected_key not in process_info:
- missing_key_set.add(expected_key)
-
- self.assertEqual(
- missing_key_set,
- set(),
- "the listed keys are missing in the qProcessInfo result")
-
- def qProcessInfo_does_not_contain_keys(self, absent_key_set):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_process_info_collection_packets()
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info response
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
-
- # Ensure the unexpected keys are not present
- unexpected_key_set = set()
- for unexpected_key in absent_key_set:
- if unexpected_key in process_info:
- unexpected_key_set.add(unexpected_key)
-
- self.assertEqual(
- unexpected_key_set,
- set(),
- "the listed keys were present but unexpected in qProcessInfo result")
-
- @skipUnlessDarwin
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qProcessInfo_contains_cputype_cpusubtype_debugserver_darwin(self):
- self.init_debugserver_test()
- self.build()
- self.qProcessInfo_contains_keys(set(['cputype', 'cpusubtype']))
-
- @skipUnlessDarwin
- @llgs_test
- def test_qProcessInfo_contains_cputype_cpusubtype_llgs_darwin(self):
- self.init_llgs_test()
- self.build()
- self.qProcessInfo_contains_keys(set(['cputype', 'cpusubtype']))
-
- @skipUnlessPlatform(["linux"])
- @llgs_test
- def test_qProcessInfo_contains_triple_llgs_linux(self):
- self.init_llgs_test()
- self.build()
- self.qProcessInfo_contains_keys(set(['triple']))
-
- @skipUnlessDarwin
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qProcessInfo_does_not_contain_triple_debugserver_darwin(self):
- self.init_debugserver_test()
- self.build()
- # We don't expect to see triple on darwin. If we do, we'll prefer triple
- # to cputype/cpusubtype and skip some darwin-based ProcessGDBRemote ArchSpec setup
- # for the remote Host and Process.
- self.qProcessInfo_does_not_contain_keys(set(['triple']))
-
- @skipUnlessDarwin
- @llgs_test
- def test_qProcessInfo_does_not_contain_triple_llgs_darwin(self):
- self.init_llgs_test()
- self.build()
- # We don't expect to see triple on darwin. If we do, we'll prefer triple
- # to cputype/cpusubtype and skip some darwin-based ProcessGDBRemote ArchSpec setup
- # for the remote Host and Process.
- self.qProcessInfo_does_not_contain_keys(set(['triple']))
-
- @skipUnlessPlatform(["linux"])
- @llgs_test
- def test_qProcessInfo_does_not_contain_cputype_cpusubtype_llgs_linux(self):
- self.init_llgs_test()
- self.build()
- self.qProcessInfo_does_not_contain_keys(set(['cputype', 'cpusubtype']))
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteRegisterState.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteRegisterState.py
deleted file mode 100644
index 2543ed6e902..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteRegisterState.py
+++ /dev/null
@@ -1,128 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteRegisterState(gdbremote_testcase.GdbRemoteTestCaseBase):
- """Test QSaveRegisterState/QRestoreRegisterState support."""
-
- mydir = TestBase.compute_mydir(__file__)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def grp_register_save_restore_works(self, with_suffix):
- # Start up the process, use thread suffix, grab main thread id.
- inferior_args = ["message:main entered", "sleep:5"]
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
-
- self.add_process_info_collection_packets()
- self.add_register_info_collection_packets()
- if with_suffix:
- self.add_thread_suffix_request_packets()
- self.add_threadinfo_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info.
- process_info = self.parse_process_info_response(context)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- # Gather register info.
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.add_lldb_register_index(reg_infos)
-
- # Pull out the register infos that we think we can bit flip
- # successfully.
- gpr_reg_infos = [
- reg_info for reg_info in reg_infos if self.is_bit_flippable_register(reg_info)]
- self.assertTrue(len(gpr_reg_infos) > 0)
-
- # Gather thread info.
- if with_suffix:
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
- thread_id = threads[0]
- self.assertIsNotNone(thread_id)
- # print("Running on thread: 0x{:x}".format(thread_id))
- else:
- thread_id = None
-
- # Save register state.
- self.reset_test_sequence()
- self.add_QSaveRegisterState_packets(thread_id)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- (success, state_id) = self.parse_QSaveRegisterState_response(context)
- self.assertTrue(success)
- self.assertIsNotNone(state_id)
- # print("saved register state id: {}".format(state_id))
-
- # Remember initial register values.
- initial_reg_values = self.read_register_values(
- gpr_reg_infos, endian, thread_id=thread_id)
- # print("initial_reg_values: {}".format(initial_reg_values))
-
- # Flip gpr register values.
- (successful_writes, failed_writes) = self.flip_all_bits_in_each_register_value(
- gpr_reg_infos, endian, thread_id=thread_id)
- # print("successful writes: {}, failed writes: {}".format(successful_writes, failed_writes))
- self.assertTrue(successful_writes > 0)
-
- flipped_reg_values = self.read_register_values(
- gpr_reg_infos, endian, thread_id=thread_id)
- # print("flipped_reg_values: {}".format(flipped_reg_values))
-
- # Restore register values.
- self.reset_test_sequence()
- self.add_QRestoreRegisterState_packets(state_id, thread_id)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify registers match initial register values.
- final_reg_values = self.read_register_values(
- gpr_reg_infos, endian, thread_id=thread_id)
- # print("final_reg_values: {}".format(final_reg_values))
- self.assertIsNotNone(final_reg_values)
- self.assertEqual(final_reg_values, initial_reg_values)
-
- @debugserver_test
- def test_grp_register_save_restore_works_with_suffix_debugserver(self):
- USE_THREAD_SUFFIX = True
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.grp_register_save_restore_works(USE_THREAD_SUFFIX)
-
- @llgs_test
- def test_grp_register_save_restore_works_with_suffix_llgs(self):
- USE_THREAD_SUFFIX = True
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.grp_register_save_restore_works(USE_THREAD_SUFFIX)
-
- @debugserver_test
- def test_grp_register_save_restore_works_no_suffix_debugserver(self):
- USE_THREAD_SUFFIX = False
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.grp_register_save_restore_works(USE_THREAD_SUFFIX)
-
- @llgs_test
- def test_grp_register_save_restore_works_no_suffix_llgs(self):
- USE_THREAD_SUFFIX = False
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.grp_register_save_restore_works(USE_THREAD_SUFFIX)
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteSingleStep.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteSingleStep.py
deleted file mode 100644
index f4e1851ff98..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteSingleStep.py
+++ /dev/null
@@ -1,41 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteSingleStep(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_single_step_only_steps_one_instruction_with_s_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=True, step_instruction="s")
-
- @llgs_test
- @expectedFailureAndroid(
- bugnumber="llvm.org/pr24739",
- archs=[
- "arm",
- "aarch64"])
- @expectedFailureAll(
- oslist=["linux"],
- archs=[
- "arm",
- "aarch64"],
- bugnumber="llvm.org/pr24739")
- @skipIf(triple='^mips')
- def test_single_step_only_steps_one_instruction_with_s_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=True, step_instruction="s")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteThreadsInStopReply.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteThreadsInStopReply.py
deleted file mode 100644
index 2bf0952e902..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemoteThreadsInStopReply.py
+++ /dev/null
@@ -1,303 +0,0 @@
-from __future__ import print_function
-
-import json
-import re
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-class TestGdbRemoteThreadsInStopReply(
- gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- ENABLE_THREADS_IN_STOP_REPLY_ENTRIES = [
- "read packet: $QListThreadsInStopReply#21",
- "send packet: $OK#00",
- ]
-
- def gather_stop_reply_fields(self, post_startup_log_lines, thread_count,
- field_names):
- # Set up the inferior args.
- inferior_args = []
- for i in range(thread_count - 1):
- inferior_args.append("thread:new")
- inferior_args.append("sleep:10")
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
-
- self.add_register_info_collection_packets()
- self.add_process_info_collection_packets()
-
- # Assumes test_sequence has anything added needed to setup the initial state.
- # (Like optionally enabling QThreadsInStopReply.)
- if post_startup_log_lines:
- self.test_sequence.add_log_lines(post_startup_log_lines, True)
- self.test_sequence.add_log_lines([
- "read packet: $c#63"
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- hw_info = self.parse_hw_info(context)
-
- # Give threads time to start up, then break.
- time.sleep(1)
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: {}".format(
- chr(3)),
- {
- "direction": "send",
- "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {
- 1: "stop_result",
- 2: "key_vals_text"}},
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Wait until all threads have started.
- threads = self.wait_for_thread_count(thread_count, timeout_seconds=3)
- self.assertIsNotNone(threads)
- self.assertEqual(len(threads), thread_count)
-
- # Run, then stop the process, grab the stop reply content.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(["read packet: $c#63",
- "read packet: {}".format(chr(3)),
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {1: "stop_result",
- 2: "key_vals_text"}},
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Parse the stop reply contents.
- key_vals_text = context.get("key_vals_text")
- self.assertIsNotNone(key_vals_text)
- kv_dict = self.parse_key_val_dict(key_vals_text)
- self.assertIsNotNone(kv_dict)
-
- result = dict();
- result["pc_register"] = hw_info["pc_register"]
- result["little_endian"] = hw_info["little_endian"]
- for key_field in field_names:
- result[key_field] = kv_dict.get(key_field)
-
- return result
-
- def gather_stop_reply_threads(self, post_startup_log_lines, thread_count):
- # Pull out threads from stop response.
- stop_reply_threads_text = self.gather_stop_reply_fields(
- post_startup_log_lines, thread_count, ["threads"])["threads"]
- if stop_reply_threads_text:
- return [int(thread_id, 16)
- for thread_id in stop_reply_threads_text.split(",")]
- else:
- return []
-
- def gather_stop_reply_pcs(self, post_startup_log_lines, thread_count):
- results = self.gather_stop_reply_fields( post_startup_log_lines,
- thread_count, ["threads", "thread-pcs"])
- if not results:
- return []
-
- threads_text = results["threads"]
- pcs_text = results["thread-pcs"]
- thread_ids = threads_text.split(",")
- pcs = pcs_text.split(",")
- self.assertTrue(len(thread_ids) == len(pcs))
-
- thread_pcs = dict()
- for i in range(0, len(pcs)):
- thread_pcs[int(thread_ids[i], 16)] = pcs[i]
-
- result = dict()
- result["thread_pcs"] = thread_pcs
- result["pc_register"] = results["pc_register"]
- result["little_endian"] = results["little_endian"]
- return result
-
- def switch_endian(self, egg):
- return "".join(reversed(re.findall("..", egg)))
-
- def parse_hw_info(self, context):
- self.assertIsNotNone(context)
- process_info = self.parse_process_info_response(context)
- endian = process_info.get("endian")
- reg_info = self.parse_register_info_packets(context)
- (pc_lldb_reg_index, pc_reg_info) = self.find_pc_reg_info(reg_info)
-
- hw_info = dict()
- hw_info["pc_register"] = pc_lldb_reg_index
- hw_info["little_endian"] = (endian == "little")
- return hw_info
-
- def gather_threads_info_pcs(self, pc_register, little_endian):
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: $jThreadsInfo#c1",
- {
- "direction": "send",
- "regex": r"^\$(.*)#[0-9a-fA-F]{2}$",
- "capture": {
- 1: "threads_info"}},
- ],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- threads_info = context.get("threads_info")
- register = str(pc_register)
- # The jThreadsInfo response is not valid JSON data, so we have to
- # clean it up first.
- jthreads_info = json.loads(re.sub(r"}]", "}", threads_info))
- thread_pcs = dict()
- for thread_info in jthreads_info:
- tid = thread_info["tid"]
- pc = thread_info["registers"][register]
- thread_pcs[tid] = self.switch_endian(pc) if little_endian else pc
-
- return thread_pcs
-
- def QListThreadsInStopReply_supported(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.test_sequence.add_log_lines(
- self.ENABLE_THREADS_IN_STOP_REPLY_ENTRIES, True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_QListThreadsInStopReply_supported_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.QListThreadsInStopReply_supported()
-
- @llgs_test
- def test_QListThreadsInStopReply_supported_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.QListThreadsInStopReply_supported()
-
- def stop_reply_reports_multiple_threads(self, thread_count):
- # Gather threads from stop notification when QThreadsInStopReply is
- # enabled.
- stop_reply_threads = self.gather_stop_reply_threads(
- self.ENABLE_THREADS_IN_STOP_REPLY_ENTRIES, thread_count)
- self.assertEqual(len(stop_reply_threads), thread_count)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_stop_reply_reports_multiple_threads_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_reports_multiple_threads(5)
-
- @llgs_test
- def test_stop_reply_reports_multiple_threads_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_reports_multiple_threads(5)
-
- def no_QListThreadsInStopReply_supplies_no_threads(self, thread_count):
- # Gather threads from stop notification when QThreadsInStopReply is not
- # enabled.
- stop_reply_threads = self.gather_stop_reply_threads(None, thread_count)
- self.assertEqual(len(stop_reply_threads), 0)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_no_QListThreadsInStopReply_supplies_no_threads_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.no_QListThreadsInStopReply_supplies_no_threads(5)
-
- @llgs_test
- def test_no_QListThreadsInStopReply_supplies_no_threads_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.no_QListThreadsInStopReply_supplies_no_threads(5)
-
- def stop_reply_reports_correct_threads(self, thread_count):
- # Gather threads from stop notification when QThreadsInStopReply is
- # enabled.
- stop_reply_threads = self.gather_stop_reply_threads(
- self.ENABLE_THREADS_IN_STOP_REPLY_ENTRIES, thread_count)
- self.assertEqual(len(stop_reply_threads), thread_count)
-
- # Gather threads from q{f,s}ThreadInfo.
- self.reset_test_sequence()
- self.add_threadinfo_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
- self.assertEqual(len(threads), thread_count)
-
- # Ensure each thread in q{f,s}ThreadInfo appears in stop reply threads
- for tid in threads:
- self.assertTrue(tid in stop_reply_threads)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_stop_reply_reports_correct_threads_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_reports_correct_threads(5)
-
- @llgs_test
- def test_stop_reply_reports_correct_threads_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_reports_correct_threads(5)
-
- def stop_reply_contains_thread_pcs(self, thread_count):
- results = self.gather_stop_reply_pcs(
- self.ENABLE_THREADS_IN_STOP_REPLY_ENTRIES, thread_count)
- stop_reply_pcs = results["thread_pcs"]
- pc_register = results["pc_register"]
- little_endian = results["little_endian"]
- self.assertEqual(len(stop_reply_pcs), thread_count)
-
- threads_info_pcs = self.gather_threads_info_pcs(pc_register,
- little_endian)
-
- self.assertEqual(len(threads_info_pcs), thread_count)
- for thread_id in stop_reply_pcs:
- self.assertTrue(thread_id in threads_info_pcs)
- self.assertTrue(int(stop_reply_pcs[thread_id], 16)
- == int(threads_info_pcs[thread_id], 16))
-
- @llgs_test
- def test_stop_reply_contains_thread_pcs_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_contains_thread_pcs(5)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @debugserver_test
- def test_stop_reply_contains_thread_pcs_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.stop_reply_contains_thread_pcs(5)
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_qThreadStopInfo.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_qThreadStopInfo.py
deleted file mode 100644
index a25484e73ba..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_qThreadStopInfo.py
+++ /dev/null
@@ -1,182 +0,0 @@
-from __future__ import print_function
-
-
-import sys
-
-import unittest2
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemote_qThreadStopInfo(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
- THREAD_COUNT = 5
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- @skipIfDarwinEmbedded # <rdar://problem/27005337>
- def gather_stop_replies_via_qThreadStopInfo(self, thread_count):
- # Set up the inferior args.
- inferior_args = []
- for i in range(thread_count - 1):
- inferior_args.append("thread:new")
- inferior_args.append("sleep:10")
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
-
- # Assumes test_sequence has anything added needed to setup the initial state.
- # (Like optionally enabling QThreadsInStopReply.)
- self.test_sequence.add_log_lines([
- "read packet: $c#63"
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Give threads time to start up, then break.
- time.sleep(1)
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: {}".format(
- chr(3)),
- {
- "direction": "send",
- "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {
- 1: "stop_result",
- 2: "key_vals_text"}},
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Wait until all threads have started.
- threads = self.wait_for_thread_count(thread_count, timeout_seconds=3)
- self.assertIsNotNone(threads)
- self.assertEqual(len(threads), thread_count)
-
- # Grab stop reply for each thread via qThreadStopInfo{tid:hex}.
- stop_replies = {}
- thread_dicts = {}
- for thread in threads:
- # Run the qThreadStopInfo command.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: $qThreadStopInfo{:x}#00".format(thread),
- {
- "direction": "send",
- "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {
- 1: "stop_result",
- 2: "key_vals_text"}},
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Parse stop reply contents.
- key_vals_text = context.get("key_vals_text")
- self.assertIsNotNone(key_vals_text)
- kv_dict = self.parse_key_val_dict(key_vals_text)
- self.assertIsNotNone(kv_dict)
-
- # Verify there is a thread and that it matches the expected thread
- # id.
- kv_thread = kv_dict.get("thread")
- self.assertIsNotNone(kv_thread)
- kv_thread_id = int(kv_thread, 16)
- self.assertEqual(kv_thread_id, thread)
-
- # Grab the stop id reported.
- stop_result_text = context.get("stop_result")
- self.assertIsNotNone(stop_result_text)
- stop_replies[kv_thread_id] = int(stop_result_text, 16)
-
- # Hang on to the key-val dictionary for the thread.
- thread_dicts[kv_thread_id] = kv_dict
-
- return (stop_replies, thread_dicts)
-
- def qThreadStopInfo_works_for_multiple_threads(self, thread_count):
- (stop_replies, _) = self.gather_stop_replies_via_qThreadStopInfo(thread_count)
- self.assertEqual(len(stop_replies), thread_count)
-
- @debugserver_test
- def test_qThreadStopInfo_works_for_multiple_threads_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_works_for_multiple_threads(self.THREAD_COUNT)
-
- @llgs_test
- def test_qThreadStopInfo_works_for_multiple_threads_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_works_for_multiple_threads(self.THREAD_COUNT)
-
- def qThreadStopInfo_only_reports_one_thread_stop_reason_during_interrupt(
- self, thread_count):
- (stop_replies, _) = self.gather_stop_replies_via_qThreadStopInfo(thread_count)
- self.assertIsNotNone(stop_replies)
-
- no_stop_reason_count = sum(
- 1 for stop_reason in list(
- stop_replies.values()) if stop_reason == 0)
- with_stop_reason_count = sum(
- 1 for stop_reason in list(
- stop_replies.values()) if stop_reason != 0)
-
- # All but one thread should report no stop reason.
- self.assertEqual(no_stop_reason_count, thread_count - 1)
-
- # Only one thread should should indicate a stop reason.
- self.assertEqual(with_stop_reason_count, 1)
-
- @debugserver_test
- def test_qThreadStopInfo_only_reports_one_thread_stop_reason_during_interrupt_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_only_reports_one_thread_stop_reason_during_interrupt(
- self.THREAD_COUNT)
-
- @llgs_test
- def test_qThreadStopInfo_only_reports_one_thread_stop_reason_during_interrupt_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_only_reports_one_thread_stop_reason_during_interrupt(
- self.THREAD_COUNT)
-
- def qThreadStopInfo_has_valid_thread_names(
- self, thread_count, expected_thread_name):
- (_, thread_dicts) = self.gather_stop_replies_via_qThreadStopInfo(thread_count)
- self.assertIsNotNone(thread_dicts)
-
- for thread_dict in list(thread_dicts.values()):
- name = thread_dict.get("name")
- self.assertIsNotNone(name)
- self.assertEqual(name, expected_thread_name)
-
- @unittest2.skip("MacOSX doesn't have a default thread name")
- @debugserver_test
- def test_qThreadStopInfo_has_valid_thread_names_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_has_valid_thread_names(self.THREAD_COUNT, "a.out")
-
- # test requires OS with set, equal thread names by default.
- @skipUnlessPlatform(["linux"])
- @llgs_test
- def test_qThreadStopInfo_has_valid_thread_names_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadStopInfo_has_valid_thread_names(self.THREAD_COUNT, "a.out")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_vCont.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_vCont.py
deleted file mode 100644
index f0380c7167e..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestGdbRemote_vCont.py
+++ /dev/null
@@ -1,159 +0,0 @@
-from __future__ import print_function
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemote_vCont(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def vCont_supports_mode(self, mode, inferior_args=None):
- # Setup the stub and set the gdb remote command stream.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
- self.add_vCont_query_packets()
-
- # Run the gdb remote command stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Pull out supported modes.
- supported_vCont_modes = self.parse_vCont_query_response(context)
- self.assertIsNotNone(supported_vCont_modes)
-
- # Verify we support the given mode.
- self.assertTrue(mode in supported_vCont_modes)
-
- def vCont_supports_c(self):
- self.vCont_supports_mode("c")
-
- def vCont_supports_C(self):
- self.vCont_supports_mode("C")
-
- def vCont_supports_s(self):
- self.vCont_supports_mode("s")
-
- def vCont_supports_S(self):
- self.vCont_supports_mode("S")
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_vCont_supports_c_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.vCont_supports_c()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @llgs_test
- def test_vCont_supports_c_llgs(self):
- self.init_llgs_test()
- self.build()
- self.vCont_supports_c()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_vCont_supports_C_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.vCont_supports_C()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @llgs_test
- def test_vCont_supports_C_llgs(self):
- self.init_llgs_test()
- self.build()
- self.vCont_supports_C()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_vCont_supports_s_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.vCont_supports_s()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @llgs_test
- def test_vCont_supports_s_llgs(self):
- self.init_llgs_test()
- self.build()
- self.vCont_supports_s()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_vCont_supports_S_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.vCont_supports_S()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @llgs_test
- def test_vCont_supports_S_llgs(self):
- self.init_llgs_test()
- self.build()
- self.vCont_supports_S()
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_single_step_only_steps_one_instruction_with_Hc_vCont_s_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=True, step_instruction="vCont;s")
-
- @llgs_test
- @expectedFailureAndroid(
- bugnumber="llvm.org/pr24739",
- archs=[
- "arm",
- "aarch64"])
- @expectedFailureAll(
- oslist=["linux"],
- archs=[
- "arm",
- "aarch64"],
- bugnumber="llvm.org/pr24739")
- @skipIf(triple='^mips')
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- def test_single_step_only_steps_one_instruction_with_Hc_vCont_s_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=True, step_instruction="vCont;s")
-
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- @debugserver_test
- def test_single_step_only_steps_one_instruction_with_vCont_s_thread_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=False, step_instruction="vCont;s:{thread}")
-
- @llgs_test
- @expectedFailureAndroid(
- bugnumber="llvm.org/pr24739",
- archs=[
- "arm",
- "aarch64"])
- @expectedFailureAll(
- oslist=["linux"],
- archs=[
- "arm",
- "aarch64"],
- bugnumber="llvm.org/pr24739")
- @skipIf(triple='^mips')
- @expectedFailureAll(oslist=["ios", "tvos", "watchos", "bridgeos"], bugnumber="rdar://27005337")
- def test_single_step_only_steps_one_instruction_with_vCont_s_thread_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.single_step_only_steps_one_instruction(
- use_Hc_packet=False, step_instruction="vCont;s:{thread}")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestLldbGdbServer.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestLldbGdbServer.py
deleted file mode 100644
index a3bad6873f1..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/TestLldbGdbServer.py
+++ /dev/null
@@ -1,1570 +0,0 @@
-"""
-Test case for testing the gdbremote protocol.
-
-Tests run against debugserver and lldb-server (llgs).
-lldb-server tests run where the lldb-server exe is
-available.
-
-This class will be broken into smaller test case classes by
-gdb remote packet functional areas. For now it contains
-the initial set of tests implemented.
-"""
-
-from __future__ import division, print_function
-
-
-import unittest2
-import gdbremote_testcase
-import lldbgdbserverutils
-import platform
-import signal
-from lldbsuite.support import seven
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test.lldbdwarf import *
-from lldbsuite.test import lldbutil
-
-
-class LldbGdbServerTestCase(gdbremote_testcase.GdbRemoteTestCaseBase, DwarfOpcodeParser):
-
- mydir = TestBase.compute_mydir(__file__)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_exe_starts_debugserver(self):
- self.init_debugserver_test()
- server = self.connect_to_debug_monitor()
-
- @llgs_test
- def test_exe_starts_llgs(self):
- self.init_llgs_test()
- server = self.connect_to_debug_monitor()
-
- def start_no_ack_mode(self):
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_start_no_ack_mode_debugserver(self):
- self.init_debugserver_test()
- self.start_no_ack_mode()
-
- @llgs_test
- def test_start_no_ack_mode_llgs(self):
- self.init_llgs_test()
- self.start_no_ack_mode()
-
- def thread_suffix_supported(self):
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.test_sequence.add_log_lines(
- ["lldb-server < 26> read packet: $QThreadSuffixSupported#e4",
- "lldb-server < 6> send packet: $OK#9a"],
- True)
-
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_thread_suffix_supported_debugserver(self):
- self.init_debugserver_test()
- self.thread_suffix_supported()
-
- @llgs_test
- def test_thread_suffix_supported_llgs(self):
- self.init_llgs_test()
- self.thread_suffix_supported()
-
- def list_threads_in_stop_reply_supported(self):
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.test_sequence.add_log_lines(
- ["lldb-server < 27> read packet: $QListThreadsInStopReply#21",
- "lldb-server < 6> send packet: $OK#9a"],
- True)
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_list_threads_in_stop_reply_supported_debugserver(self):
- self.init_debugserver_test()
- self.list_threads_in_stop_reply_supported()
-
- @llgs_test
- def test_list_threads_in_stop_reply_supported_llgs(self):
- self.init_llgs_test()
- self.list_threads_in_stop_reply_supported()
-
- def c_packet_works(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(
- ["read packet: $c#63",
- "send packet: $W00#00"],
- True)
-
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_c_packet_works_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.c_packet_works()
-
- @llgs_test
- def test_c_packet_works_llgs(self):
- self.init_llgs_test()
- self.build()
- self.c_packet_works()
-
- def inferior_print_exit(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # build launch args
- launch_args += ["hello, world"]
-
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8",
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"hello, world\r\n")},
- "send packet: $W00#00"],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_inferior_print_exit_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.inferior_print_exit()
-
- @llgs_test
- @expectedFlakeyLinux("llvm.org/pr25652")
- def test_inferior_print_exit_llgs(self):
- self.init_llgs_test()
- self.build()
- self.inferior_print_exit()
-
- def first_launch_stop_reply_thread_matches_first_qC(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # build launch args
- launch_args += ["hello, world"]
-
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(["read packet: $qC#00",
- {"direction": "send",
- "regex": r"^\$QC([0-9a-fA-F]+)#",
- "capture": {1: "thread_id"}},
- "read packet: $?#00",
- {"direction": "send",
- "regex": r"^\$T[0-9a-fA-F]{2}thread:([0-9a-fA-F]+)",
- "expect_captures": {1: "thread_id"}}],
- True)
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_first_launch_stop_reply_thread_matches_first_qC_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.first_launch_stop_reply_thread_matches_first_qC()
-
- @llgs_test
- def test_first_launch_stop_reply_thread_matches_first_qC_llgs(self):
- self.init_llgs_test()
- self.build()
- self.first_launch_stop_reply_thread_matches_first_qC()
-
- def attach_commandline_continue_app_exits(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8",
- "send packet: $W00#00"],
- True)
- self.expect_gdbremote_sequence()
-
- # Wait a moment for completed and now-detached inferior process to
- # clear.
- time.sleep(1)
-
- if not lldb.remote_platform:
- # Process should be dead now. Reap results.
- poll_result = procs["inferior"].poll()
- self.assertIsNotNone(poll_result)
-
- # Where possible, verify at the system level that the process is not
- # running.
- self.assertFalse(
- lldbgdbserverutils.process_is_running(
- procs["inferior"].pid, False))
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_attach_commandline_continue_app_exits_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_continue_app_exits()
-
- @llgs_test
- def test_attach_commandline_continue_app_exits_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.attach_commandline_continue_app_exits()
-
- def qRegisterInfo_returns_one_valid_result(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.test_sequence.add_log_lines(
- ["read packet: $qRegisterInfo0#00",
- {"direction": "send", "regex": r"^\$(.+);#[0-9A-Fa-f]{2}", "capture": {1: "reginfo_0"}}],
- True)
-
- # Run the stream
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- reg_info_packet = context.get("reginfo_0")
- self.assertIsNotNone(reg_info_packet)
- self.assert_valid_reg_info(
- lldbgdbserverutils.parse_reg_info_response(reg_info_packet))
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qRegisterInfo_returns_one_valid_result_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.qRegisterInfo_returns_one_valid_result()
-
- @llgs_test
- def test_qRegisterInfo_returns_one_valid_result_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qRegisterInfo_returns_one_valid_result()
-
- def qRegisterInfo_returns_all_valid_results(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream.
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.add_register_info_collection_packets()
-
- # Run the stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Validate that each register info returned validates.
- for reg_info in self.parse_register_info_packets(context):
- self.assert_valid_reg_info(reg_info)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qRegisterInfo_returns_all_valid_results_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.qRegisterInfo_returns_all_valid_results()
-
- @llgs_test
- def test_qRegisterInfo_returns_all_valid_results_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qRegisterInfo_returns_all_valid_results()
-
- def qRegisterInfo_contains_required_generics(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.add_register_info_collection_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather register info entries.
- reg_infos = self.parse_register_info_packets(context)
-
- # Collect all generic registers found.
- generic_regs = {
- reg_info['generic']: 1 for reg_info in reg_infos if 'generic' in reg_info}
-
- # Ensure we have a program counter register.
- self.assertTrue('pc' in generic_regs)
-
- # Ensure we have a frame pointer register. PPC64le's FP is the same as SP
- if self.getArchitecture() != 'powerpc64le':
- self.assertTrue('fp' in generic_regs)
-
- # Ensure we have a stack pointer register.
- self.assertTrue('sp' in generic_regs)
-
- # Ensure we have a flags register.
- self.assertTrue('flags' in generic_regs)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qRegisterInfo_contains_required_generics_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.qRegisterInfo_contains_required_generics()
-
- @llgs_test
- def test_qRegisterInfo_contains_required_generics_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qRegisterInfo_contains_required_generics()
-
- def qRegisterInfo_contains_at_least_one_register_set(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.add_register_info_collection_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather register info entries.
- reg_infos = self.parse_register_info_packets(context)
-
- # Collect all register sets found.
- register_sets = {
- reg_info['set']: 1 for reg_info in reg_infos if 'set' in reg_info}
- self.assertTrue(len(register_sets) >= 1)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qRegisterInfo_contains_at_least_one_register_set_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.qRegisterInfo_contains_at_least_one_register_set()
-
- @llgs_test
- def test_qRegisterInfo_contains_at_least_one_register_set_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qRegisterInfo_contains_at_least_one_register_set()
-
- def targetHasAVX(self):
- triple = self.dbg.GetSelectedPlatform().GetTriple()
-
- # TODO other platforms, please implement this function
- if not re.match(".*-.*-linux", triple):
- return True
-
- # Need to do something different for non-Linux/Android targets
- if lldb.remote_platform:
- self.runCmd('platform get-file "/proc/cpuinfo" "cpuinfo"')
- cpuinfo_path = "cpuinfo"
- self.addTearDownHook(lambda: os.unlink("cpuinfo"))
- else:
- cpuinfo_path = "/proc/cpuinfo"
-
- f = open(cpuinfo_path, 'r')
- cpuinfo = f.read()
- f.close()
- return " avx " in cpuinfo
-
- def qRegisterInfo_contains_avx_registers(self):
- launch_args = self.install_and_create_launch_args()
-
- server = self.connect_to_debug_monitor()
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream
- self.add_no_ack_remote_stream()
- self.add_verified_launch_packets(launch_args)
- self.add_register_info_collection_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather register info entries.
- reg_infos = self.parse_register_info_packets(context)
-
- # Collect all generics found.
- register_sets = {
- reg_info['set']: 1 for reg_info in reg_infos if 'set' in reg_info}
- self.assertEqual(
- self.targetHasAVX(),
- "Advanced Vector Extensions" in register_sets)
-
- @llgs_test
- def test_qRegisterInfo_contains_avx_registers_llgs(self):
- self.init_llgs_test()
- self.build()
- self.qRegisterInfo_contains_avx_registers()
-
- def qThreadInfo_contains_thread(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_threadinfo_collection_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather threadinfo entries.
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
-
- # We should have exactly one thread.
- self.assertEqual(len(threads), 1)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qThreadInfo_contains_thread_launch_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadInfo_contains_thread()
-
- @llgs_test
- def test_qThreadInfo_contains_thread_launch_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadInfo_contains_thread()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qThreadInfo_contains_thread_attach_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.qThreadInfo_contains_thread()
-
- @llgs_test
- def test_qThreadInfo_contains_thread_attach_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.qThreadInfo_contains_thread()
-
- def qThreadInfo_matches_qC(self):
- procs = self.prep_debug_monitor_and_inferior()
-
- self.add_threadinfo_collection_packets()
- self.test_sequence.add_log_lines(
- ["read packet: $qC#00",
- {"direction": "send", "regex": r"^\$QC([0-9a-fA-F]+)#", "capture": {1: "thread_id"}}
- ], True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather threadinfo entries.
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
-
- # We should have exactly one thread from threadinfo.
- self.assertEqual(len(threads), 1)
-
- # We should have a valid thread_id from $QC.
- QC_thread_id_hex = context.get("thread_id")
- self.assertIsNotNone(QC_thread_id_hex)
- QC_thread_id = int(QC_thread_id_hex, 16)
-
- # Those two should be the same.
- self.assertEqual(threads[0], QC_thread_id)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qThreadInfo_matches_qC_launch_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadInfo_matches_qC()
-
- @llgs_test
- def test_qThreadInfo_matches_qC_launch_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qThreadInfo_matches_qC()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qThreadInfo_matches_qC_attach_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.qThreadInfo_matches_qC()
-
- @llgs_test
- def test_qThreadInfo_matches_qC_attach_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.qThreadInfo_matches_qC()
-
- def p_returns_correct_data_size_for_each_qRegisterInfo(self):
- procs = self.prep_debug_monitor_and_inferior()
- self.add_register_info_collection_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather register info entries.
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.assertTrue(len(reg_infos) > 0)
-
- byte_order = self.get_target_byte_order()
-
- # Read value for each register.
- reg_index = 0
- for reg_info in reg_infos:
- # Skip registers that don't have a register set. For x86, these are
- # the DRx registers, which have no LLDB-kind register number and thus
- # cannot be read via normal
- # NativeRegisterContext::ReadRegister(reg_info,...) calls.
- if not "set" in reg_info:
- continue
-
- # Clear existing packet expectations.
- self.reset_test_sequence()
-
- # Run the register query
- self.test_sequence.add_log_lines(
- ["read packet: $p{0:x}#00".format(reg_index),
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify the response length.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
-
- if "dynamic_size_dwarf_expr_bytes" in reg_info:
- self.updateRegInfoBitsize(reg_info, byte_order)
- self.assertEqual(len(p_response), 2 * int(reg_info["bitsize"]) / 8)
-
- # Increment loop
- reg_index += 1
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_p_returns_correct_data_size_for_each_qRegisterInfo_launch_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.p_returns_correct_data_size_for_each_qRegisterInfo()
-
- @llgs_test
- def test_p_returns_correct_data_size_for_each_qRegisterInfo_launch_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.p_returns_correct_data_size_for_each_qRegisterInfo()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_p_returns_correct_data_size_for_each_qRegisterInfo_attach_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.p_returns_correct_data_size_for_each_qRegisterInfo()
-
- @llgs_test
- def test_p_returns_correct_data_size_for_each_qRegisterInfo_attach_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.p_returns_correct_data_size_for_each_qRegisterInfo()
-
- def Hg_switches_to_3_threads(self):
- # Startup the inferior with three threads (main + 2 new ones).
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["thread:new", "thread:new"])
-
- # Let the inferior process have a few moments to start up the thread
- # when launched. (The launch scenario has no time to run, so threads
- # won't be there yet.)
- self.run_process_then_stop(run_seconds=1)
-
- # Wait at most x seconds for 3 threads to be present.
- threads = self.wait_for_thread_count(3, timeout_seconds=5)
- self.assertEqual(len(threads), 3)
-
- # verify we can $H to each thead, and $qC matches the thread we set.
- for thread in threads:
- # Change to each thread, verify current thread id.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $Hg{0:x}#00".format(thread), # Set current thread.
- "send packet: $OK#00",
- "read packet: $qC#00",
- {"direction": "send", "regex": r"^\$QC([0-9a-fA-F]+)#", "capture": {1: "thread_id"}}],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify the thread id.
- self.assertIsNotNone(context.get("thread_id"))
- self.assertEqual(int(context.get("thread_id"), 16), thread)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_Hg_switches_to_3_threads_launch_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.Hg_switches_to_3_threads()
-
- @llgs_test
- def test_Hg_switches_to_3_threads_launch_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.Hg_switches_to_3_threads()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_Hg_switches_to_3_threads_attach_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_attach()
- self.Hg_switches_to_3_threads()
-
- @llgs_test
- def test_Hg_switches_to_3_threads_attach_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_attach()
- self.Hg_switches_to_3_threads()
-
- def Hc_then_Csignal_signals_correct_thread(self, segfault_signo):
- # NOTE only run this one in inferior-launched mode: we can't grab inferior stdout when running attached,
- # and the test requires getting stdout from the exe.
-
- NUM_THREADS = 3
-
- # Startup the inferior with three threads (main + NUM_THREADS-1 worker threads).
- # inferior_args=["thread:print-ids"]
- inferior_args = ["thread:segfault"]
- for i in range(NUM_THREADS - 1):
- # if i > 0:
- # Give time between thread creation/segfaulting for the handler to work.
- # inferior_args.append("sleep:1")
- inferior_args.append("thread:new")
- inferior_args.append("sleep:10")
-
- # Launch/attach. (In our case, this should only ever be launched since
- # we need inferior stdout/stderr).
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=inferior_args)
- self.test_sequence.add_log_lines(["read packet: $c#63"], True)
- context = self.expect_gdbremote_sequence()
-
- # Let the inferior process have a few moments to start up the thread when launched.
- # context = self.run_process_then_stop(run_seconds=1)
-
- # Wait at most x seconds for all threads to be present.
- # threads = self.wait_for_thread_count(NUM_THREADS, timeout_seconds=5)
- # self.assertEquals(len(threads), NUM_THREADS)
-
- signaled_tids = {}
- print_thread_ids = {}
-
- # Switch to each thread, deliver a signal, and verify signal delivery
- for i in range(NUM_THREADS - 1):
- # Run until SIGSEGV comes in.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines([{"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);",
- "capture": {1: "signo",
- 2: "thread_id"}}],
- True)
-
- context = self.expect_gdbremote_sequence(timeout_seconds=10)
- self.assertIsNotNone(context)
- signo = context.get("signo")
- self.assertEqual(int(signo, 16), segfault_signo)
-
- # Ensure we haven't seen this tid yet.
- thread_id = int(context.get("thread_id"), 16)
- self.assertFalse(thread_id in signaled_tids)
- signaled_tids[thread_id] = 1
-
- # Send SIGUSR1 to the thread that signaled the SIGSEGV.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- # Set the continue thread.
- # Set current thread.
- "read packet: $Hc{0:x}#00".format(thread_id),
- "send packet: $OK#00",
-
- # Continue sending the signal number to the continue thread.
- # The commented out packet is a way to do this same operation without using
- # a $Hc (but this test is testing $Hc, so we'll stick with the former).
- "read packet: $C{0:x}#00".format(lldbutil.get_signal_number('SIGUSR1')),
- # "read packet: $vCont;C{0:x}:{1:x};c#00".format(lldbutil.get_signal_number('SIGUSR1'), thread_id),
-
- # FIXME: Linux does not report the thread stop on the delivered signal (SIGUSR1 here). MacOSX debugserver does.
- # But MacOSX debugserver isn't guaranteeing the thread the signal handler runs on, so currently its an XFAIL.
- # Need to rectify behavior here. The linux behavior is more intuitive to me since we're essentially swapping out
- # an about-to-be-delivered signal (for which we already sent a stop packet) to a different signal.
- # {"direction":"send", "regex":r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture":{1:"stop_signo", 2:"stop_thread_id"} },
- # "read packet: $c#63",
- {"type": "output_match", "regex": r"^received SIGUSR1 on thread id: ([0-9a-fA-F]+)\r\nthread ([0-9a-fA-F]+): past SIGSEGV\r\n", "capture": {1: "print_thread_id", 2: "post_handle_thread_id"}},
- ],
- True)
-
- # Run the sequence.
- context = self.expect_gdbremote_sequence(timeout_seconds=10)
- self.assertIsNotNone(context)
-
- # Ensure the stop signal is the signal we delivered.
- # stop_signo = context.get("stop_signo")
- # self.assertIsNotNone(stop_signo)
- # self.assertEquals(int(stop_signo,16), lldbutil.get_signal_number('SIGUSR1'))
-
- # Ensure the stop thread is the thread to which we delivered the signal.
- # stop_thread_id = context.get("stop_thread_id")
- # self.assertIsNotNone(stop_thread_id)
- # self.assertEquals(int(stop_thread_id,16), thread_id)
-
- # Ensure we haven't seen this thread id yet. The inferior's
- # self-obtained thread ids are not guaranteed to match the stub
- # tids (at least on MacOSX).
- print_thread_id = context.get("print_thread_id")
- self.assertIsNotNone(print_thread_id)
- print_thread_id = int(print_thread_id, 16)
- self.assertFalse(print_thread_id in print_thread_ids)
-
- # Now remember this print (i.e. inferior-reflected) thread id and
- # ensure we don't hit it again.
- print_thread_ids[print_thread_id] = 1
-
- # Ensure post signal-handle thread id matches the thread that
- # initially raised the SIGSEGV.
- post_handle_thread_id = context.get("post_handle_thread_id")
- self.assertIsNotNone(post_handle_thread_id)
- post_handle_thread_id = int(post_handle_thread_id, 16)
- self.assertEqual(post_handle_thread_id, print_thread_id)
-
- @unittest2.expectedFailure()
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_Hc_then_Csignal_signals_correct_thread_launch_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- # Darwin debugserver translates some signals like SIGSEGV into some gdb
- # expectations about fixed signal numbers.
- self.Hc_then_Csignal_signals_correct_thread(self.TARGET_EXC_BAD_ACCESS)
-
- @llgs_test
- def test_Hc_then_Csignal_signals_correct_thread_launch_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.Hc_then_Csignal_signals_correct_thread(
- lldbutil.get_signal_number('SIGSEGV'))
-
- def m_packet_reads_memory(self):
- # This is the memory we will write into the inferior and then ensure we
- # can read back with $m.
- MEMORY_CONTENTS = "Test contents 0123456789 ABCDEFGHIJKLMNOPQRSTUVWXYZ abcdefghijklmnopqrstuvwxyz"
-
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=[
- "set-message:%s" %
- MEMORY_CONTENTS,
- "get-data-address-hex:g_message",
- "sleep:5"])
-
- # Run the process
- self.test_sequence.add_log_lines(
- [
- # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the message buffer within the inferior.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"data address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "message_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the message address.
- self.assertIsNotNone(context.get("message_address"))
- message_address = int(context.get("message_address"), 16)
-
- # Grab contents from the inferior.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $m{0:x},{1:x}#00".format(message_address, len(MEMORY_CONTENTS)),
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$", "capture": {1: "read_contents"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Ensure what we read from inferior memory is what we wrote.
- self.assertIsNotNone(context.get("read_contents"))
- read_contents = seven.unhexlify(context.get("read_contents"))
- self.assertEqual(read_contents, MEMORY_CONTENTS)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_m_packet_reads_memory_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.m_packet_reads_memory()
-
- @llgs_test
- def test_m_packet_reads_memory_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.m_packet_reads_memory()
-
- def qMemoryRegionInfo_is_supported(self):
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior()
-
- # Ask if it supports $qMemoryRegionInfo.
- self.test_sequence.add_log_lines(
- ["read packet: $qMemoryRegionInfo#00",
- "send packet: $OK#00"
- ], True)
- self.expect_gdbremote_sequence()
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qMemoryRegionInfo_is_supported_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_is_supported()
-
- @llgs_test
- def test_qMemoryRegionInfo_is_supported_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_is_supported()
-
- def qMemoryRegionInfo_reports_code_address_as_executable(self):
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["get-code-address-hex:hello", "sleep:5"])
-
- # Run the process
- self.test_sequence.add_log_lines(
- [
- # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the message buffer within the inferior.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"code address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "code_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the code address.
- self.assertIsNotNone(context.get("code_address"))
- code_address = int(context.get("code_address"), 16)
-
- # Grab memory region info from the inferior.
- self.reset_test_sequence()
- self.add_query_memory_region_packets(code_address)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- mem_region_dict = self.parse_memory_region_packet(context)
-
- # Ensure there are no errors reported.
- self.assertFalse("error" in mem_region_dict)
-
- # Ensure code address is readable and executable.
- self.assertTrue("permissions" in mem_region_dict)
- self.assertTrue("r" in mem_region_dict["permissions"])
- self.assertTrue("x" in mem_region_dict["permissions"])
-
- # Ensure the start address and size encompass the address we queried.
- self.assert_address_within_memory_region(code_address, mem_region_dict)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qMemoryRegionInfo_reports_code_address_as_executable_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_code_address_as_executable()
-
- @llgs_test
- def test_qMemoryRegionInfo_reports_code_address_as_executable_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_code_address_as_executable()
-
- def qMemoryRegionInfo_reports_stack_address_as_readable_writeable(self):
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["get-stack-address-hex:", "sleep:5"])
-
- # Run the process
- self.test_sequence.add_log_lines(
- [
- # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the message buffer within the inferior.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"stack address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "stack_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the address.
- self.assertIsNotNone(context.get("stack_address"))
- stack_address = int(context.get("stack_address"), 16)
-
- # Grab memory region info from the inferior.
- self.reset_test_sequence()
- self.add_query_memory_region_packets(stack_address)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- mem_region_dict = self.parse_memory_region_packet(context)
-
- # Ensure there are no errors reported.
- self.assertFalse("error" in mem_region_dict)
-
- # Ensure address is readable and executable.
- self.assertTrue("permissions" in mem_region_dict)
- self.assertTrue("r" in mem_region_dict["permissions"])
- self.assertTrue("w" in mem_region_dict["permissions"])
-
- # Ensure the start address and size encompass the address we queried.
- self.assert_address_within_memory_region(
- stack_address, mem_region_dict)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qMemoryRegionInfo_reports_stack_address_as_readable_writeable_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_stack_address_as_readable_writeable()
-
- @llgs_test
- def test_qMemoryRegionInfo_reports_stack_address_as_readable_writeable_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_stack_address_as_readable_writeable()
-
- def qMemoryRegionInfo_reports_heap_address_as_readable_writeable(self):
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["get-heap-address-hex:", "sleep:5"])
-
- # Run the process
- self.test_sequence.add_log_lines(
- [
- # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the message buffer within the inferior.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"heap address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "heap_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the address.
- self.assertIsNotNone(context.get("heap_address"))
- heap_address = int(context.get("heap_address"), 16)
-
- # Grab memory region info from the inferior.
- self.reset_test_sequence()
- self.add_query_memory_region_packets(heap_address)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- mem_region_dict = self.parse_memory_region_packet(context)
-
- # Ensure there are no errors reported.
- self.assertFalse("error" in mem_region_dict)
-
- # Ensure address is readable and executable.
- self.assertTrue("permissions" in mem_region_dict)
- self.assertTrue("r" in mem_region_dict["permissions"])
- self.assertTrue("w" in mem_region_dict["permissions"])
-
- # Ensure the start address and size encompass the address we queried.
- self.assert_address_within_memory_region(heap_address, mem_region_dict)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qMemoryRegionInfo_reports_heap_address_as_readable_writeable_debugserver(
- self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_heap_address_as_readable_writeable()
-
- @llgs_test
- def test_qMemoryRegionInfo_reports_heap_address_as_readable_writeable_llgs(
- self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qMemoryRegionInfo_reports_heap_address_as_readable_writeable()
-
- def breakpoint_set_and_remove_work(self, want_hardware=False):
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=[
- "get-code-address-hex:hello",
- "sleep:1",
- "call-function:hello"])
-
- # Run the process
- self.add_register_info_collection_packets()
- self.add_process_info_collection_packets()
- self.test_sequence.add_log_lines(
- [ # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the function call entry point.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"code address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "function_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather process info - we need endian of target to handle register
- # value conversions.
- process_info = self.parse_process_info_response(context)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- # Gather register info entries.
- reg_infos = self.parse_register_info_packets(context)
- (pc_lldb_reg_index, pc_reg_info) = self.find_pc_reg_info(reg_infos)
- self.assertIsNotNone(pc_lldb_reg_index)
- self.assertIsNotNone(pc_reg_info)
-
- # Grab the function address.
- self.assertIsNotNone(context.get("function_address"))
- function_address = int(context.get("function_address"), 16)
-
- # Get current target architecture
- target_arch = self.getArchitecture()
-
- # Set the breakpoint.
- if (target_arch == "arm") or (target_arch == "aarch64"):
- # TODO: Handle case when setting breakpoint in thumb code
- BREAKPOINT_KIND = 4
- else:
- BREAKPOINT_KIND = 1
-
- # Set default packet type to Z0 (software breakpoint)
- z_packet_type = 0
-
- # If hardware breakpoint is requested set packet type to Z1
- if want_hardware == True:
- z_packet_type = 1
-
- self.reset_test_sequence()
- self.add_set_breakpoint_packets(
- function_address,
- z_packet_type,
- do_continue=True,
- breakpoint_kind=BREAKPOINT_KIND)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify the stop signal reported was the breakpoint signal number.
- stop_signo = context.get("stop_signo")
- self.assertIsNotNone(stop_signo)
- self.assertEqual(int(stop_signo, 16),
- lldbutil.get_signal_number('SIGTRAP'))
-
- # Ensure we did not receive any output. If the breakpoint was not set, we would
- # see output (from a launched process with captured stdio) printing a hello, world message.
- # That would indicate the breakpoint didn't take.
- self.assertEqual(len(context["O_content"]), 0)
-
- # Verify that the PC for the main thread is where we expect it - right at the breakpoint address.
- # This acts as a another validation on the register reading code.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- # Print the PC. This should match the breakpoint address.
- "read packet: $p{0:x}#00".format(pc_lldb_reg_index),
- # Capture $p results.
- {"direction": "send",
- "regex": r"^\$([0-9a-fA-F]+)#",
- "capture": {1: "p_response"}},
- ], True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify the PC is where we expect. Note response is in endianness of
- # the inferior.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
-
- # Convert from target endian to int.
- returned_pc = lldbgdbserverutils.unpack_register_hex_unsigned(
- endian, p_response)
- self.assertEqual(returned_pc, function_address)
-
- # Verify that a breakpoint remove and continue gets us the expected
- # output.
- self.reset_test_sequence()
-
- # Add breakpoint remove packets
- self.add_remove_breakpoint_packets(
- function_address,
- z_packet_type,
- breakpoint_kind=BREAKPOINT_KIND)
-
- self.test_sequence.add_log_lines(
- [
- # Continue running.
- "read packet: $c#63",
- # We should now receive the output from the call.
- {"type": "output_match", "regex": r"^hello, world\r\n$"},
- # And wait for program completion.
- {"direction": "send", "regex": r"^\$W00(.*)#[0-9a-fA-F]{2}$"},
- ], True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_software_breakpoint_set_and_remove_work_debugserver(self):
- self.init_debugserver_test()
- if self.getArchitecture() == "arm":
- # TODO: Handle case when setting breakpoint in thumb code
- self.build(dictionary={'CFLAGS_EXTRAS': '-marm'})
- else:
- self.build()
- self.set_inferior_startup_launch()
- self.breakpoint_set_and_remove_work(want_hardware=False)
-
- @llgs_test
- @expectedFlakeyLinux("llvm.org/pr25652")
- def test_software_breakpoint_set_and_remove_work_llgs(self):
- self.init_llgs_test()
- if self.getArchitecture() == "arm":
- # TODO: Handle case when setting breakpoint in thumb code
- self.build(dictionary={'CFLAGS_EXTRAS': '-marm'})
- else:
- self.build()
- self.set_inferior_startup_launch()
- self.breakpoint_set_and_remove_work(want_hardware=False)
-
- @debugserver_test
- @skipUnlessPlatform(oslist=['linux'])
- @expectedFailureAndroid
- @skipIf(archs=no_match(['arm', 'aarch64']))
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_hardware_breakpoint_set_and_remove_work_debugserver(self):
- self.init_debugserver_test()
- if self.getArchitecture() == "arm":
- # TODO: Handle case when setting breakpoint in thumb code
- self.build(dictionary={'CFLAGS_EXTRAS': '-marm'})
- else:
- self.build()
- self.set_inferior_startup_launch()
- self.breakpoint_set_and_remove_work(want_hardware=True)
-
- @llgs_test
- @skipUnlessPlatform(oslist=['linux'])
- @expectedFailureAndroid
- @skipIf(archs=no_match(['arm', 'aarch64']))
- def test_hardware_breakpoint_set_and_remove_work_llgs(self):
- self.init_llgs_test()
- if self.getArchitecture() == "arm":
- # TODO: Handle case when setting breakpoint in thumb code
- self.build(dictionary={'CFLAGS_EXTRAS': '-marm'})
- else:
- self.build()
- self.set_inferior_startup_launch()
- self.breakpoint_set_and_remove_work(want_hardware=True)
-
- def qSupported_returns_known_stub_features(self):
- # Start up the stub and start/prep the inferior.
- procs = self.prep_debug_monitor_and_inferior()
- self.add_qSupported_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Retrieve the qSupported features.
- supported_dict = self.parse_qSupported_response(context)
- self.assertIsNotNone(supported_dict)
- self.assertTrue(len(supported_dict) > 0)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_qSupported_returns_known_stub_features_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qSupported_returns_known_stub_features()
-
- @llgs_test
- def test_qSupported_returns_known_stub_features_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.qSupported_returns_known_stub_features()
-
- def written_M_content_reads_back_correctly(self):
- TEST_MESSAGE = "Hello, memory"
-
- # Start up the stub and start/prep the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=[
- "set-message:xxxxxxxxxxxxxX",
- "get-data-address-hex:g_message",
- "sleep:1",
- "print-message:"])
- self.test_sequence.add_log_lines(
- [
- # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the message buffer within the inferior.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": self.maybe_strict_output_regex(r"data address: 0x([0-9a-fA-F]+)\r\n"),
- "capture": {1: "message_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the message address.
- self.assertIsNotNone(context.get("message_address"))
- message_address = int(context.get("message_address"), 16)
-
- # Hex-encode the test message, adding null termination.
- hex_encoded_message = seven.hexlify(TEST_MESSAGE)
-
- # Write the message to the inferior. Verify that we can read it with the hex-encoded (m)
- # and binary (x) memory read packets.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $M{0:x},{1:x}:{2}#00".format(message_address, len(TEST_MESSAGE), hex_encoded_message),
- "send packet: $OK#00",
- "read packet: $m{0:x},{1:x}#00".format(message_address, len(TEST_MESSAGE)),
- "send packet: ${0}#00".format(hex_encoded_message),
- "read packet: $x{0:x},{1:x}#00".format(message_address, len(TEST_MESSAGE)),
- "send packet: ${0}#00".format(TEST_MESSAGE),
- "read packet: $m{0:x},4#00".format(message_address),
- "send packet: ${0}#00".format(hex_encoded_message[0:8]),
- "read packet: $x{0:x},4#00".format(message_address),
- "send packet: ${0}#00".format(TEST_MESSAGE[0:4]),
- "read packet: $c#63",
- {"type": "output_match", "regex": r"^message: (.+)\r\n$", "capture": {1: "printed_message"}},
- "send packet: $W00#00",
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Ensure what we read from inferior memory is what we wrote.
- printed_message = context.get("printed_message")
- self.assertIsNotNone(printed_message)
- self.assertEqual(printed_message, TEST_MESSAGE + "X")
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_written_M_content_reads_back_correctly_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.written_M_content_reads_back_correctly()
-
- @llgs_test
- @expectedFlakeyLinux("llvm.org/pr25652")
- def test_written_M_content_reads_back_correctly_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.written_M_content_reads_back_correctly()
-
- def P_writes_all_gpr_registers(self):
- # Start inferior debug session, grab all register info.
- procs = self.prep_debug_monitor_and_inferior(inferior_args=["sleep:2"])
- self.add_register_info_collection_packets()
- self.add_process_info_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Process register infos.
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.add_lldb_register_index(reg_infos)
-
- # Process endian.
- process_info = self.parse_process_info_response(context)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- # Pull out the register infos that we think we can bit flip
- # successfully,.
- gpr_reg_infos = [
- reg_info for reg_info in reg_infos if self.is_bit_flippable_register(reg_info)]
- self.assertTrue(len(gpr_reg_infos) > 0)
-
- # Write flipped bit pattern of existing value to each register.
- (successful_writes, failed_writes) = self.flip_all_bits_in_each_register_value(
- gpr_reg_infos, endian)
- # print("successful writes: {}, failed writes: {}".format(successful_writes, failed_writes))
- self.assertTrue(successful_writes > 0)
-
- # Note: as of this moment, a hefty number of the GPR writes are failing with E32 (everything except rax-rdx, rdi, rsi, rbp).
- # Come back to this. I have the test rigged to verify that at least some
- # of the bit-flip writes work.
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_P_writes_all_gpr_registers_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.P_writes_all_gpr_registers()
-
- @llgs_test
- def test_P_writes_all_gpr_registers_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.P_writes_all_gpr_registers()
-
- def P_and_p_thread_suffix_work(self):
- # Startup the inferior with three threads.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["thread:new", "thread:new"])
- self.add_thread_suffix_request_packets()
- self.add_register_info_collection_packets()
- self.add_process_info_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- process_info = self.parse_process_info_response(context)
- self.assertIsNotNone(process_info)
- endian = process_info.get("endian")
- self.assertIsNotNone(endian)
-
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.add_lldb_register_index(reg_infos)
-
- reg_index = self.select_modifiable_register(reg_infos)
- self.assertIsNotNone(reg_index)
- reg_byte_size = int(reg_infos[reg_index]["bitsize"]) // 8
- self.assertTrue(reg_byte_size > 0)
-
- # Run the process a bit so threads can start up, and collect register
- # info.
- context = self.run_process_then_stop(run_seconds=1)
- self.assertIsNotNone(context)
-
- # Wait for 3 threads to be present.
- threads = self.wait_for_thread_count(3, timeout_seconds=5)
- self.assertEqual(len(threads), 3)
-
- expected_reg_values = []
- register_increment = 1
- next_value = None
-
- # Set the same register in each of 3 threads to a different value.
- # Verify each one has the unique value.
- for thread in threads:
- # If we don't have a next value yet, start it with the initial read
- # value + 1
- if not next_value:
- # Read pre-existing register value.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $p{0:x};thread:{1:x}#00".format(reg_index, thread),
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Set the next value to use for writing as the increment plus
- # current value.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
- next_value = lldbgdbserverutils.unpack_register_hex_unsigned(
- endian, p_response)
-
- # Set new value using P and thread suffix.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: $P{0:x}={1};thread:{2:x}#00".format(
- reg_index,
- lldbgdbserverutils.pack_register_hex(
- endian,
- next_value,
- byte_size=reg_byte_size),
- thread),
- "send packet: $OK#00",
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Save the value we set.
- expected_reg_values.append(next_value)
-
- # Increment value for next thread to use (we want them all
- # different so we can verify they wrote to each thread correctly
- # next.)
- next_value += register_increment
-
- # Revisit each thread and verify they have the expected value set for
- # the register we wrote.
- thread_index = 0
- for thread in threads:
- # Read pre-existing register value.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $p{0:x};thread:{1:x}#00".format(reg_index, thread),
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Get the register value.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
- read_value = lldbgdbserverutils.unpack_register_hex_unsigned(
- endian, p_response)
-
- # Make sure we read back what we wrote.
- self.assertEqual(read_value, expected_reg_values[thread_index])
- thread_index += 1
-
- # Note: as of this moment, a hefty number of the GPR writes are failing
- # with E32 (everything except rax-rdx, rdi, rsi, rbp).
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_P_and_p_thread_suffix_work_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.set_inferior_startup_launch()
- self.P_and_p_thread_suffix_work()
-
- @llgs_test
- def test_P_and_p_thread_suffix_work_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.P_and_p_thread_suffix_work()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubReverseConnect.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubReverseConnect.py
deleted file mode 100644
index 102f2e344af..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubReverseConnect.py
+++ /dev/null
@@ -1,98 +0,0 @@
-from __future__ import print_function
-
-import gdbremote_testcase
-import lldbgdbserverutils
-import re
-import select
-import socket
-import time
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestStubReverseConnect(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- _DEFAULT_TIMEOUT = 20
-
- def setUp(self):
- # Set up the test.
- gdbremote_testcase.GdbRemoteTestCaseBase.setUp(self)
-
- # Create a listener on a local port.
- self.listener_socket = self.create_listener_socket()
- self.assertIsNotNone(self.listener_socket)
- self.listener_port = self.listener_socket.getsockname()[1]
-
- def create_listener_socket(self, timeout_seconds=_DEFAULT_TIMEOUT):
- sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
- self.assertIsNotNone(sock)
-
- sock.settimeout(timeout_seconds)
- sock.bind(("127.0.0.1", 0))
- sock.listen(1)
-
- def tear_down_listener():
- try:
- sock.shutdown(socket.SHUT_RDWR)
- except:
- # ignore
- None
-
- self.addTearDownHook(tear_down_listener)
- return sock
-
- def reverse_connect_works(self):
- # Indicate stub startup should do a reverse connect.
- appended_stub_args = ["--reverse-connect"]
- if self.debug_monitor_extra_args:
- self.debug_monitor_extra_args += appended_stub_args
- else:
- self.debug_monitor_extra_args = appended_stub_args
-
- self.stub_hostname = "127.0.0.1"
- self.port = self.listener_port
-
- triple = self.dbg.GetSelectedPlatform().GetTriple()
- if re.match(".*-.*-.*-android", triple):
- self.forward_adb_port(
- self.port,
- self.port,
- "reverse",
- self.stub_device)
-
- # Start the stub.
- server = self.launch_debug_monitor(logfile=sys.stdout)
- self.assertIsNotNone(server)
- self.assertTrue(
- lldbgdbserverutils.process_is_running(
- server.pid, True))
-
- # Listen for the stub's connection to us.
- (stub_socket, address) = self.listener_socket.accept()
- self.assertIsNotNone(stub_socket)
- self.assertIsNotNone(address)
- print("connected to stub {} on {}".format(
- address, stub_socket.getsockname()))
-
- # Verify we can do the handshake. If that works, we'll call it good.
- self.do_handshake(stub_socket, timeout_seconds=self._DEFAULT_TIMEOUT)
-
- # Clean up.
- stub_socket.shutdown(socket.SHUT_RDWR)
-
- @debugserver_test
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def test_reverse_connect_works_debugserver(self):
- self.init_debugserver_test(use_named_pipe=False)
- self.set_inferior_startup_launch()
- self.reverse_connect_works()
-
- @llgs_test
- @skipIfRemote # reverse connect is not a supported use case for now
- def test_reverse_connect_works_llgs(self):
- self.init_llgs_test(use_named_pipe=False)
- self.set_inferior_startup_launch()
- self.reverse_connect_works()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubSetSID.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubSetSID.py
deleted file mode 100644
index 27daf30f189..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/commandline/TestStubSetSID.py
+++ /dev/null
@@ -1,86 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import lldbgdbserverutils
-import os
-import select
-import tempfile
-import time
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestStubSetSIDTestCase(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def get_stub_sid(self, extra_stub_args=None):
- # Launch debugserver
- if extra_stub_args:
- self.debug_monitor_extra_args += extra_stub_args
-
- server = self.launch_debug_monitor()
- self.assertIsNotNone(server)
- self.assertTrue(
- lldbgdbserverutils.process_is_running(
- server.pid, True))
-
- # Get the process id for the stub.
- return os.getsid(server.pid)
-
- def sid_is_same_without_setsid(self):
- stub_sid = self.get_stub_sid()
- self.assertEqual(stub_sid, os.getsid(0))
-
- def sid_is_different_with_setsid(self):
- stub_sid = self.get_stub_sid(["--setsid"])
- self.assertNotEqual(stub_sid, os.getsid(0))
-
- def sid_is_different_with_S(self):
- stub_sid = self.get_stub_sid(["-S"])
- self.assertNotEqual(stub_sid, os.getsid(0))
-
- @debugserver_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- def test_sid_is_same_without_setsid_debugserver(self):
- self.init_debugserver_test()
- self.set_inferior_startup_launch()
- self.sid_is_same_without_setsid()
-
- @llgs_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- @expectedFailureAll(oslist=['freebsd'])
- def test_sid_is_same_without_setsid_llgs(self):
- self.init_llgs_test()
- self.set_inferior_startup_launch()
- self.sid_is_same_without_setsid()
-
- @debugserver_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- def test_sid_is_different_with_setsid_debugserver(self):
- self.init_debugserver_test()
- self.set_inferior_startup_launch()
- self.sid_is_different_with_setsid()
-
- @llgs_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- def test_sid_is_different_with_setsid_llgs(self):
- self.init_llgs_test()
- self.set_inferior_startup_launch()
- self.sid_is_different_with_setsid()
-
- @debugserver_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- def test_sid_is_different_with_S_debugserver(self):
- self.init_debugserver_test()
- self.set_inferior_startup_launch()
- self.sid_is_different_with_S()
-
- @llgs_test
- @skipIfRemote # --setsid not used on remote platform and currently it is also impossible to get the sid of lldb-platform running on a remote target
- def test_sid_is_different_with_S_llgs(self):
- self.init_llgs_test()
- self.set_inferior_startup_launch()
- self.sid_is_different_with_S()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/gdbremote_testcase.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/gdbremote_testcase.py
deleted file mode 100644
index e7c63bf21e8..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/gdbremote_testcase.py
+++ /dev/null
@@ -1,1643 +0,0 @@
-"""
-Base class for gdb-remote test cases.
-"""
-
-from __future__ import division, print_function
-
-
-import errno
-import os
-import os.path
-import platform
-import random
-import re
-import select
-import signal
-import socket
-import subprocess
-import sys
-import tempfile
-import time
-from lldbsuite.test import configuration
-from lldbsuite.test.lldbtest import *
-from lldbsuite.support import seven
-from lldbgdbserverutils import *
-import logging
-
-
-class _ConnectionRefused(IOError):
- pass
-
-
-class GdbRemoteTestCaseBase(TestBase):
-
- NO_DEBUG_INFO_TESTCASE = True
-
- _TIMEOUT_SECONDS = 120
-
- _GDBREMOTE_KILL_PACKET = "$k#6b"
-
- # Start the inferior separately, attach to the inferior on the stub
- # command line.
- _STARTUP_ATTACH = "attach"
- # Start the inferior separately, start the stub without attaching, allow
- # the test to attach to the inferior however it wants (e.g. $vAttach;pid).
- _STARTUP_ATTACH_MANUALLY = "attach_manually"
- # Start the stub, and launch the inferior with an $A packet via the
- # initial packet stream.
- _STARTUP_LAUNCH = "launch"
-
- # GDB Signal numbers that are not target-specific used for common
- # exceptions
- TARGET_EXC_BAD_ACCESS = 0x91
- TARGET_EXC_BAD_INSTRUCTION = 0x92
- TARGET_EXC_ARITHMETIC = 0x93
- TARGET_EXC_EMULATION = 0x94
- TARGET_EXC_SOFTWARE = 0x95
- TARGET_EXC_BREAKPOINT = 0x96
-
- _verbose_log_handler = None
- _log_formatter = logging.Formatter(
- fmt='%(asctime)-15s %(levelname)-8s %(message)s')
-
- def setUpBaseLogging(self):
- self.logger = logging.getLogger(__name__)
-
- if len(self.logger.handlers) > 0:
- return # We have set up this handler already
-
- self.logger.propagate = False
- self.logger.setLevel(logging.DEBUG)
-
- # log all warnings to stderr
- handler = logging.StreamHandler()
- handler.setLevel(logging.WARNING)
- handler.setFormatter(self._log_formatter)
- self.logger.addHandler(handler)
-
- def isVerboseLoggingRequested(self):
- # We will report our detailed logs if the user requested that the "gdb-remote" channel is
- # logged.
- return any(("gdb-remote" in channel)
- for channel in lldbtest_config.channels)
-
- def setUp(self):
- TestBase.setUp(self)
-
- self.setUpBaseLogging()
- self.debug_monitor_extra_args = []
- self._pump_queues = socket_packet_pump.PumpQueues()
-
- if self.isVerboseLoggingRequested():
- # If requested, full logs go to a log file
- self._verbose_log_handler = logging.FileHandler(
- self.log_basename + "-host.log")
- self._verbose_log_handler.setFormatter(self._log_formatter)
- self._verbose_log_handler.setLevel(logging.DEBUG)
- self.logger.addHandler(self._verbose_log_handler)
-
- self.test_sequence = GdbRemoteTestSequence(self.logger)
- self.set_inferior_startup_launch()
- self.port = self.get_next_port()
- self.named_pipe_path = None
- self.named_pipe = None
- self.named_pipe_fd = None
- self.stub_sends_two_stop_notifications_on_kill = False
- if configuration.lldb_platform_url:
- if configuration.lldb_platform_url.startswith('unix-'):
- url_pattern = '(.+)://\[?(.+?)\]?/.*'
- else:
- url_pattern = '(.+)://(.+):\d+'
- scheme, host = re.match(
- url_pattern, configuration.lldb_platform_url).groups()
- if configuration.lldb_platform_name == 'remote-android' and host != 'localhost':
- self.stub_device = host
- self.stub_hostname = 'localhost'
- else:
- self.stub_device = None
- self.stub_hostname = host
- else:
- self.stub_hostname = "localhost"
-
- def tearDown(self):
- self._pump_queues.verify_queues_empty()
-
- self.logger.removeHandler(self._verbose_log_handler)
- self._verbose_log_handler = None
- TestBase.tearDown(self)
-
- def getLocalServerLogFile(self):
- return self.log_basename + "-server.log"
-
- def setUpServerLogging(self, is_llgs):
- if len(lldbtest_config.channels) == 0:
- return # No logging requested
-
- if lldb.remote_platform:
- log_file = lldbutil.join_remote_paths(
- lldb.remote_platform.GetWorkingDirectory(), "server.log")
- else:
- log_file = self.getLocalServerLogFile()
-
- if is_llgs:
- self.debug_monitor_extra_args.append("--log-file=" + log_file)
- self.debug_monitor_extra_args.append(
- "--log-channels={}".format(":".join(lldbtest_config.channels)))
- else:
- self.debug_monitor_extra_args = [
- "--log-file=" + log_file, "--log-flags=0x800000"]
-
- def get_next_port(self):
- return 12000 + random.randint(0, 3999)
-
- def reset_test_sequence(self):
- self.test_sequence = GdbRemoteTestSequence(self.logger)
-
- def create_named_pipe(self):
- # Create a temp dir and name for a pipe.
- temp_dir = tempfile.mkdtemp()
- named_pipe_path = os.path.join(temp_dir, "stub_port_number")
-
- # Create the named pipe.
- os.mkfifo(named_pipe_path)
-
- # Open the read side of the pipe in non-blocking mode. This will
- # return right away, ready or not.
- named_pipe_fd = os.open(named_pipe_path, os.O_RDONLY | os.O_NONBLOCK)
-
- # Create the file for the named pipe. Note this will follow semantics of
- # a non-blocking read side of a named pipe, which has different semantics
- # than a named pipe opened for read in non-blocking mode.
- named_pipe = os.fdopen(named_pipe_fd, "r")
- self.assertIsNotNone(named_pipe)
-
- def shutdown_named_pipe():
- # Close the pipe.
- try:
- named_pipe.close()
- except:
- print("failed to close named pipe")
- None
-
- # Delete the pipe.
- try:
- os.remove(named_pipe_path)
- except:
- print("failed to delete named pipe: {}".format(named_pipe_path))
- None
-
- # Delete the temp directory.
- try:
- os.rmdir(temp_dir)
- except:
- print(
- "failed to delete temp dir: {}, directory contents: '{}'".format(
- temp_dir, os.listdir(temp_dir)))
- None
-
- # Add the shutdown hook to clean up the named pipe.
- self.addTearDownHook(shutdown_named_pipe)
-
- # Clear the port so the stub selects a port number.
- self.port = 0
-
- return (named_pipe_path, named_pipe, named_pipe_fd)
-
- def get_stub_port_from_named_socket(self, read_timeout_seconds=5):
- # Wait for something to read with a max timeout.
- (ready_readers, _, _) = select.select(
- [self.named_pipe_fd], [], [], read_timeout_seconds)
- self.assertIsNotNone(
- ready_readers,
- "write side of pipe has not written anything - stub isn't writing to pipe.")
- self.assertNotEqual(
- len(ready_readers),
- 0,
- "write side of pipe has not written anything - stub isn't writing to pipe.")
-
- # Read the port from the named pipe.
- stub_port_raw = self.named_pipe.read()
- self.assertIsNotNone(stub_port_raw)
- self.assertNotEqual(
- len(stub_port_raw),
- 0,
- "no content to read on pipe")
-
- # Trim null byte, convert to int.
- stub_port_raw = stub_port_raw[:-1]
- stub_port = int(stub_port_raw)
- self.assertTrue(stub_port > 0)
-
- return stub_port
-
- def init_llgs_test(self, use_named_pipe=True):
- if lldb.remote_platform:
- # Remote platforms don't support named pipe based port negotiation
- use_named_pipe = False
-
- # Grab the ppid from /proc/[shell pid]/stat
- err, retcode, shell_stat = self.run_platform_command(
- "cat /proc/$$/stat")
- self.assertTrue(
- err.Success() and retcode == 0,
- "Failed to read file /proc/$$/stat: %s, retcode: %d" %
- (err.GetCString(),
- retcode))
-
- # [pid] ([executable]) [state] [*ppid*]
- pid = re.match(r"^\d+ \(.+\) . (\d+)", shell_stat).group(1)
- err, retcode, ls_output = self.run_platform_command(
- "ls -l /proc/%s/exe" % pid)
- self.assertTrue(
- err.Success() and retcode == 0,
- "Failed to read file /proc/%s/exe: %s, retcode: %d" %
- (pid,
- err.GetCString(),
- retcode))
- exe = ls_output.split()[-1]
-
- # If the binary has been deleted, the link name has " (deleted)" appended.
- # Remove if it's there.
- self.debug_monitor_exe = re.sub(r' \(deleted\)$', '', exe)
- else:
- self.debug_monitor_exe = get_lldb_server_exe()
- if not self.debug_monitor_exe:
- self.skipTest("lldb-server exe not found")
-
- self.debug_monitor_extra_args = ["gdbserver"]
- self.setUpServerLogging(is_llgs=True)
-
- if use_named_pipe:
- (self.named_pipe_path, self.named_pipe,
- self.named_pipe_fd) = self.create_named_pipe()
-
- def init_debugserver_test(self, use_named_pipe=True):
- self.debug_monitor_exe = get_debugserver_exe()
- if not self.debug_monitor_exe:
- self.skipTest("debugserver exe not found")
- self.setUpServerLogging(is_llgs=False)
- if use_named_pipe:
- (self.named_pipe_path, self.named_pipe,
- self.named_pipe_fd) = self.create_named_pipe()
- # The debugserver stub has a race on handling the 'k' command, so it sends an X09 right away, then sends the real X notification
- # when the process truly dies.
- self.stub_sends_two_stop_notifications_on_kill = True
-
- def forward_adb_port(self, source, target, direction, device):
- adb = ['adb'] + (['-s', device] if device else []) + [direction]
-
- def remove_port_forward():
- subprocess.call(adb + ["--remove", "tcp:%d" % source])
-
- subprocess.call(adb + ["tcp:%d" % source, "tcp:%d" % target])
- self.addTearDownHook(remove_port_forward)
-
- def _verify_socket(self, sock):
- # Normally, when the remote stub is not ready, we will get ECONNREFUSED during the
- # connect() attempt. However, due to the way how ADB forwarding works, on android targets
- # the connect() will always be successful, but the connection will be immediately dropped
- # if ADB could not connect on the remote side. This function tries to detect this
- # situation, and report it as "connection refused" so that the upper layers attempt the
- # connection again.
- triple = self.dbg.GetSelectedPlatform().GetTriple()
- if not re.match(".*-.*-.*-android", triple):
- return # Not android.
- can_read, _, _ = select.select([sock], [], [], 0.1)
- if sock not in can_read:
- return # Data is not available, but the connection is alive.
- if len(sock.recv(1, socket.MSG_PEEK)) == 0:
- raise _ConnectionRefused() # Got EOF, connection dropped.
-
- def create_socket(self):
- sock = socket.socket()
- logger = self.logger
-
- triple = self.dbg.GetSelectedPlatform().GetTriple()
- if re.match(".*-.*-.*-android", triple):
- self.forward_adb_port(
- self.port,
- self.port,
- "forward",
- self.stub_device)
-
- logger.info(
- "Connecting to debug monitor on %s:%d",
- self.stub_hostname,
- self.port)
- connect_info = (self.stub_hostname, self.port)
- try:
- sock.connect(connect_info)
- except socket.error as serr:
- if serr.errno == errno.ECONNREFUSED:
- raise _ConnectionRefused()
- raise serr
-
- def shutdown_socket():
- if sock:
- try:
- # send the kill packet so lldb-server shuts down gracefully
- sock.sendall(GdbRemoteTestCaseBase._GDBREMOTE_KILL_PACKET)
- except:
- logger.warning(
- "failed to send kill packet to debug monitor: {}; ignoring".format(
- sys.exc_info()[0]))
-
- try:
- sock.close()
- except:
- logger.warning(
- "failed to close socket to debug monitor: {}; ignoring".format(
- sys.exc_info()[0]))
-
- self.addTearDownHook(shutdown_socket)
-
- self._verify_socket(sock)
-
- return sock
-
- def set_inferior_startup_launch(self):
- self._inferior_startup = self._STARTUP_LAUNCH
-
- def set_inferior_startup_attach(self):
- self._inferior_startup = self._STARTUP_ATTACH
-
- def set_inferior_startup_attach_manually(self):
- self._inferior_startup = self._STARTUP_ATTACH_MANUALLY
-
- def get_debug_monitor_command_line_args(self, attach_pid=None):
- if lldb.remote_platform:
- commandline_args = self.debug_monitor_extra_args + \
- ["*:{}".format(self.port)]
- else:
- commandline_args = self.debug_monitor_extra_args + \
- ["127.0.0.1:{}".format(self.port)]
-
- if attach_pid:
- commandline_args += ["--attach=%d" % attach_pid]
- if self.named_pipe_path:
- commandline_args += ["--named-pipe", self.named_pipe_path]
- return commandline_args
-
- def get_target_byte_order(self):
- inferior_exe_path = self.getBuildArtifact("a.out")
- target = self.dbg.CreateTarget(inferior_exe_path)
- return target.GetByteOrder()
-
- def launch_debug_monitor(self, attach_pid=None, logfile=None):
- # Create the command line.
- commandline_args = self.get_debug_monitor_command_line_args(
- attach_pid=attach_pid)
-
- # Start the server.
- server = self.spawnSubprocess(
- self.debug_monitor_exe,
- commandline_args,
- install_remote=False)
- self.addTearDownHook(self.cleanupSubprocesses)
- self.assertIsNotNone(server)
-
- # If we're receiving the stub's listening port from the named pipe, do
- # that here.
- if self.named_pipe:
- self.port = self.get_stub_port_from_named_socket()
-
- return server
-
- def connect_to_debug_monitor(self, attach_pid=None):
- if self.named_pipe:
- # Create the stub.
- server = self.launch_debug_monitor(attach_pid=attach_pid)
- self.assertIsNotNone(server)
-
- def shutdown_debug_monitor():
- try:
- server.terminate()
- except:
- logger.warning(
- "failed to terminate server for debug monitor: {}; ignoring".format(
- sys.exc_info()[0]))
- self.addTearDownHook(shutdown_debug_monitor)
-
- # Schedule debug monitor to be shut down during teardown.
- logger = self.logger
-
- # Attach to the stub and return a socket opened to it.
- self.sock = self.create_socket()
- return server
-
- # We're using a random port algorithm to try not to collide with other ports,
- # and retry a max # times.
- attempts = 0
- MAX_ATTEMPTS = 20
-
- while attempts < MAX_ATTEMPTS:
- server = self.launch_debug_monitor(attach_pid=attach_pid)
-
- # Schedule debug monitor to be shut down during teardown.
- logger = self.logger
-
- def shutdown_debug_monitor():
- try:
- server.terminate()
- except:
- logger.warning(
- "failed to terminate server for debug monitor: {}; ignoring".format(
- sys.exc_info()[0]))
- self.addTearDownHook(shutdown_debug_monitor)
-
- connect_attemps = 0
- MAX_CONNECT_ATTEMPTS = 10
-
- while connect_attemps < MAX_CONNECT_ATTEMPTS:
- # Create a socket to talk to the server
- try:
- logger.info("Connect attempt %d", connect_attemps + 1)
- self.sock = self.create_socket()
- return server
- except _ConnectionRefused as serr:
- # Ignore, and try again.
- pass
- time.sleep(0.5)
- connect_attemps += 1
-
- # We should close the server here to be safe.
- server.terminate()
-
- # Increment attempts.
- print(
- "connect to debug monitor on port %d failed, attempt #%d of %d" %
- (self.port, attempts + 1, MAX_ATTEMPTS))
- attempts += 1
-
- # And wait a random length of time before next attempt, to avoid
- # collisions.
- time.sleep(random.randint(1, 5))
-
- # Now grab a new port number.
- self.port = self.get_next_port()
-
- raise Exception(
- "failed to create a socket to the launched debug monitor after %d tries" %
- attempts)
-
- def launch_process_for_attach(
- self,
- inferior_args=None,
- sleep_seconds=3,
- exe_path=None):
- # We're going to start a child process that the debug monitor stub can later attach to.
- # This process needs to be started so that it just hangs around for a while. We'll
- # have it sleep.
- if not exe_path:
- exe_path = self.getBuildArtifact("a.out")
-
- args = []
- if inferior_args:
- args.extend(inferior_args)
- if sleep_seconds:
- args.append("sleep:%d" % sleep_seconds)
-
- inferior = self.spawnSubprocess(exe_path, args)
-
- def shutdown_process_for_attach():
- try:
- inferior.terminate()
- except:
- logger.warning(
- "failed to terminate inferior process for attach: {}; ignoring".format(
- sys.exc_info()[0]))
- self.addTearDownHook(shutdown_process_for_attach)
- return inferior
-
- def prep_debug_monitor_and_inferior(
- self,
- inferior_args=None,
- inferior_sleep_seconds=3,
- inferior_exe_path=None):
- """Prep the debug monitor, the inferior, and the expected packet stream.
-
- Handle the separate cases of using the debug monitor in attach-to-inferior mode
- and in launch-inferior mode.
-
- For attach-to-inferior mode, the inferior process is first started, then
- the debug monitor is started in attach to pid mode (using --attach on the
- stub command line), and the no-ack-mode setup is appended to the packet
- stream. The packet stream is not yet executed, ready to have more expected
- packet entries added to it.
-
- For launch-inferior mode, the stub is first started, then no ack mode is
- setup on the expected packet stream, then the verified launch packets are added
- to the expected socket stream. The packet stream is not yet executed, ready
- to have more expected packet entries added to it.
-
- The return value is:
- {inferior:<inferior>, server:<server>}
- """
- inferior = None
- attach_pid = None
-
- if self._inferior_startup == self._STARTUP_ATTACH or self._inferior_startup == self._STARTUP_ATTACH_MANUALLY:
- # Launch the process that we'll use as the inferior.
- inferior = self.launch_process_for_attach(
- inferior_args=inferior_args,
- sleep_seconds=inferior_sleep_seconds,
- exe_path=inferior_exe_path)
- self.assertIsNotNone(inferior)
- self.assertTrue(inferior.pid > 0)
- if self._inferior_startup == self._STARTUP_ATTACH:
- # In this case, we want the stub to attach via the command
- # line, so set the command line attach pid here.
- attach_pid = inferior.pid
-
- if self._inferior_startup == self._STARTUP_LAUNCH:
- # Build launch args
- if not inferior_exe_path:
- inferior_exe_path = self.getBuildArtifact("a.out")
-
- if lldb.remote_platform:
- remote_path = lldbutil.append_to_process_working_directory(self,
- os.path.basename(inferior_exe_path))
- remote_file_spec = lldb.SBFileSpec(remote_path, False)
- err = lldb.remote_platform.Install(lldb.SBFileSpec(
- inferior_exe_path, True), remote_file_spec)
- if err.Fail():
- raise Exception(
- "remote_platform.Install('%s', '%s') failed: %s" %
- (inferior_exe_path, remote_path, err))
- inferior_exe_path = remote_path
-
- launch_args = [inferior_exe_path]
- if inferior_args:
- launch_args.extend(inferior_args)
-
- # Launch the debug monitor stub, attaching to the inferior.
- server = self.connect_to_debug_monitor(attach_pid=attach_pid)
- self.assertIsNotNone(server)
-
- # Build the expected protocol stream
- self.add_no_ack_remote_stream()
- if self._inferior_startup == self._STARTUP_LAUNCH:
- self.add_verified_launch_packets(launch_args)
-
- return {"inferior": inferior, "server": server}
-
- def expect_socket_recv(
- self,
- sock,
- expected_content_regex,
- timeout_seconds):
- response = ""
- timeout_time = time.time() + timeout_seconds
-
- while not expected_content_regex.match(
- response) and time.time() < timeout_time:
- can_read, _, _ = select.select([sock], [], [], timeout_seconds)
- if can_read and sock in can_read:
- recv_bytes = sock.recv(4096)
- if recv_bytes:
- response += seven.bitcast_to_string(recv_bytes)
-
- self.assertTrue(expected_content_regex.match(response))
-
- def expect_socket_send(self, sock, content, timeout_seconds):
- request_bytes_remaining = content
- timeout_time = time.time() + timeout_seconds
-
- while len(request_bytes_remaining) > 0 and time.time() < timeout_time:
- _, can_write, _ = select.select([], [sock], [], timeout_seconds)
- if can_write and sock in can_write:
- written_byte_count = sock.send(request_bytes_remaining.encode())
- request_bytes_remaining = request_bytes_remaining[
- written_byte_count:]
- self.assertEqual(len(request_bytes_remaining), 0)
-
- def do_handshake(self, stub_socket, timeout_seconds=5):
- # Write the ack.
- self.expect_socket_send(stub_socket, "+", timeout_seconds)
-
- # Send the start no ack mode packet.
- NO_ACK_MODE_REQUEST = "$QStartNoAckMode#b0"
- bytes_sent = stub_socket.send(NO_ACK_MODE_REQUEST.encode())
- self.assertEqual(bytes_sent, len(NO_ACK_MODE_REQUEST))
-
- # Receive the ack and "OK"
- self.expect_socket_recv(stub_socket, re.compile(
- r"^\+\$OK#[0-9a-fA-F]{2}$"), timeout_seconds)
-
- # Send the final ack.
- self.expect_socket_send(stub_socket, "+", timeout_seconds)
-
- def add_no_ack_remote_stream(self):
- self.test_sequence.add_log_lines(
- ["read packet: +",
- "read packet: $QStartNoAckMode#b0",
- "send packet: +",
- "send packet: $OK#9a",
- "read packet: +"],
- True)
-
- def add_verified_launch_packets(self, launch_args):
- self.test_sequence.add_log_lines(
- ["read packet: %s" % build_gdbremote_A_packet(launch_args),
- "send packet: $OK#00",
- "read packet: $qLaunchSuccess#a5",
- "send packet: $OK#00"],
- True)
-
- def add_thread_suffix_request_packets(self):
- self.test_sequence.add_log_lines(
- ["read packet: $QThreadSuffixSupported#e4",
- "send packet: $OK#00",
- ], True)
-
- def add_process_info_collection_packets(self):
- self.test_sequence.add_log_lines(
- ["read packet: $qProcessInfo#dc",
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$", "capture": {1: "process_info_raw"}}],
- True)
-
- _KNOWN_PROCESS_INFO_KEYS = [
- "pid",
- "parent-pid",
- "real-uid",
- "real-gid",
- "effective-uid",
- "effective-gid",
- "cputype",
- "cpusubtype",
- "ostype",
- "triple",
- "vendor",
- "endian",
- "elf_abi",
- "ptrsize"
- ]
-
- def parse_process_info_response(self, context):
- # Ensure we have a process info response.
- self.assertIsNotNone(context)
- process_info_raw = context.get("process_info_raw")
- self.assertIsNotNone(process_info_raw)
-
- # Pull out key:value; pairs.
- process_info_dict = {
- match.group(1): match.group(2) for match in re.finditer(
- r"([^:]+):([^;]+);", process_info_raw)}
-
- # Validate keys are known.
- for (key, val) in list(process_info_dict.items()):
- self.assertTrue(key in self._KNOWN_PROCESS_INFO_KEYS)
- self.assertIsNotNone(val)
-
- return process_info_dict
-
- def add_register_info_collection_packets(self):
- self.test_sequence.add_log_lines(
- [{"type": "multi_response", "query": "qRegisterInfo", "append_iteration_suffix": True,
- "end_regex": re.compile(r"^\$(E\d+)?#[0-9a-fA-F]{2}$"),
- "save_key": "reg_info_responses"}],
- True)
-
- def parse_register_info_packets(self, context):
- """Return an array of register info dictionaries, one per register info."""
- reg_info_responses = context.get("reg_info_responses")
- self.assertIsNotNone(reg_info_responses)
-
- # Parse register infos.
- return [parse_reg_info_response(reg_info_response)
- for reg_info_response in reg_info_responses]
-
- def expect_gdbremote_sequence(self, timeout_seconds=None):
- if not timeout_seconds:
- timeout_seconds = self._TIMEOUT_SECONDS
- return expect_lldb_gdbserver_replay(
- self,
- self.sock,
- self.test_sequence,
- self._pump_queues,
- timeout_seconds,
- self.logger)
-
- _KNOWN_REGINFO_KEYS = [
- "name",
- "alt-name",
- "bitsize",
- "offset",
- "encoding",
- "format",
- "set",
- "gcc",
- "ehframe",
- "dwarf",
- "generic",
- "container-regs",
- "invalidate-regs",
- "dynamic_size_dwarf_expr_bytes",
- "dynamic_size_dwarf_len"
- ]
-
- def assert_valid_reg_info(self, reg_info):
- # Assert we know about all the reginfo keys parsed.
- for key in reg_info:
- self.assertTrue(key in self._KNOWN_REGINFO_KEYS)
-
- # Check the bare-minimum expected set of register info keys.
- self.assertTrue("name" in reg_info)
- self.assertTrue("bitsize" in reg_info)
- self.assertTrue("offset" in reg_info)
- self.assertTrue("encoding" in reg_info)
- self.assertTrue("format" in reg_info)
-
- def find_pc_reg_info(self, reg_infos):
- lldb_reg_index = 0
- for reg_info in reg_infos:
- if ("generic" in reg_info) and (reg_info["generic"] == "pc"):
- return (lldb_reg_index, reg_info)
- lldb_reg_index += 1
-
- return (None, None)
-
- def add_lldb_register_index(self, reg_infos):
- """Add a "lldb_register_index" key containing the 0-baed index of each reg_infos entry.
-
- We'll use this when we want to call packets like P/p with a register index but do so
- on only a subset of the full register info set.
- """
- self.assertIsNotNone(reg_infos)
-
- reg_index = 0
- for reg_info in reg_infos:
- reg_info["lldb_register_index"] = reg_index
- reg_index += 1
-
- def add_query_memory_region_packets(self, address):
- self.test_sequence.add_log_lines(
- ["read packet: $qMemoryRegionInfo:{0:x}#00".format(address),
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$", "capture": {1: "memory_region_response"}}],
- True)
-
- def parse_key_val_dict(self, key_val_text, allow_dupes=True):
- self.assertIsNotNone(key_val_text)
- kv_dict = {}
- for match in re.finditer(r";?([^:]+):([^;]+)", key_val_text):
- key = match.group(1)
- val = match.group(2)
- if key in kv_dict:
- if allow_dupes:
- if isinstance(kv_dict[key], list):
- kv_dict[key].append(val)
- else:
- # Promote to list
- kv_dict[key] = [kv_dict[key], val]
- else:
- self.fail(
- "key '{}' already present when attempting to add value '{}' (text='{}', dict={})".format(
- key, val, key_val_text, kv_dict))
- else:
- kv_dict[key] = val
- return kv_dict
-
- def parse_memory_region_packet(self, context):
- # Ensure we have a context.
- self.assertIsNotNone(context.get("memory_region_response"))
-
- # Pull out key:value; pairs.
- mem_region_dict = self.parse_key_val_dict(
- context.get("memory_region_response"))
-
- # Validate keys are known.
- for (key, val) in list(mem_region_dict.items()):
- self.assertTrue(
- key in [
- "start",
- "size",
- "permissions",
- "name",
- "error"])
- self.assertIsNotNone(val)
-
- # Return the dictionary of key-value pairs for the memory region.
- return mem_region_dict
-
- def assert_address_within_memory_region(
- self, test_address, mem_region_dict):
- self.assertIsNotNone(mem_region_dict)
- self.assertTrue("start" in mem_region_dict)
- self.assertTrue("size" in mem_region_dict)
-
- range_start = int(mem_region_dict["start"], 16)
- range_size = int(mem_region_dict["size"], 16)
- range_end = range_start + range_size
-
- if test_address < range_start:
- self.fail(
- "address 0x{0:x} comes before range 0x{1:x} - 0x{2:x} (size 0x{3:x})".format(
- test_address,
- range_start,
- range_end,
- range_size))
- elif test_address >= range_end:
- self.fail(
- "address 0x{0:x} comes after range 0x{1:x} - 0x{2:x} (size 0x{3:x})".format(
- test_address,
- range_start,
- range_end,
- range_size))
-
- def add_threadinfo_collection_packets(self):
- self.test_sequence.add_log_lines(
- [{"type": "multi_response", "first_query": "qfThreadInfo", "next_query": "qsThreadInfo",
- "append_iteration_suffix": False, "end_regex": re.compile(r"^\$(l)?#[0-9a-fA-F]{2}$"),
- "save_key": "threadinfo_responses"}],
- True)
-
- def parse_threadinfo_packets(self, context):
- """Return an array of thread ids (decimal ints), one per thread."""
- threadinfo_responses = context.get("threadinfo_responses")
- self.assertIsNotNone(threadinfo_responses)
-
- thread_ids = []
- for threadinfo_response in threadinfo_responses:
- new_thread_infos = parse_threadinfo_response(threadinfo_response)
- thread_ids.extend(new_thread_infos)
- return thread_ids
-
- def wait_for_thread_count(self, thread_count, timeout_seconds=3):
- start_time = time.time()
- timeout_time = start_time + timeout_seconds
-
- actual_thread_count = 0
- while actual_thread_count < thread_count:
- self.reset_test_sequence()
- self.add_threadinfo_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
-
- actual_thread_count = len(threads)
-
- if time.time() > timeout_time:
- raise Exception(
- 'timed out after {} seconds while waiting for theads: waiting for at least {} threads, found {}'.format(
- timeout_seconds, thread_count, actual_thread_count))
-
- return threads
-
- def add_set_breakpoint_packets(
- self,
- address,
- z_packet_type=0,
- do_continue=True,
- breakpoint_kind=1):
- self.test_sequence.add_log_lines(
- [ # Set the breakpoint.
- "read packet: $Z{2},{0:x},{1}#00".format(
- address, breakpoint_kind, z_packet_type),
- # Verify the stub could set it.
- "send packet: $OK#00",
- ], True)
-
- if (do_continue):
- self.test_sequence.add_log_lines(
- [ # Continue the inferior.
- "read packet: $c#63",
- # Expect a breakpoint stop report.
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);",
- "capture": {1: "stop_signo",
- 2: "stop_thread_id"}},
- ], True)
-
- def add_remove_breakpoint_packets(
- self,
- address,
- z_packet_type=0,
- breakpoint_kind=1):
- self.test_sequence.add_log_lines(
- [ # Remove the breakpoint.
- "read packet: $z{2},{0:x},{1}#00".format(
- address, breakpoint_kind, z_packet_type),
- # Verify the stub could unset it.
- "send packet: $OK#00",
- ], True)
-
- def add_qSupported_packets(self):
- self.test_sequence.add_log_lines(
- ["read packet: $qSupported#00",
- {"direction": "send", "regex": r"^\$(.*)#[0-9a-fA-F]{2}", "capture": {1: "qSupported_response"}},
- ], True)
-
- _KNOWN_QSUPPORTED_STUB_FEATURES = [
- "augmented-libraries-svr4-read",
- "PacketSize",
- "QStartNoAckMode",
- "QThreadSuffixSupported",
- "QListThreadsInStopReply",
- "qXfer:auxv:read",
- "qXfer:libraries:read",
- "qXfer:libraries-svr4:read",
- "qXfer:features:read",
- "qEcho",
- "QPassSignals"
- ]
-
- def parse_qSupported_response(self, context):
- self.assertIsNotNone(context)
-
- raw_response = context.get("qSupported_response")
- self.assertIsNotNone(raw_response)
-
- # For values with key=val, the dict key and vals are set as expected. For feature+, feature- and feature?, the
- # +,-,? is stripped from the key and set as the value.
- supported_dict = {}
- for match in re.finditer(r";?([^=;]+)(=([^;]+))?", raw_response):
- key = match.group(1)
- val = match.group(3)
-
- # key=val: store as is
- if val and len(val) > 0:
- supported_dict[key] = val
- else:
- if len(key) < 2:
- raise Exception(
- "singular stub feature is too short: must be stub_feature{+,-,?}")
- supported_type = key[-1]
- key = key[:-1]
- if not supported_type in ["+", "-", "?"]:
- raise Exception(
- "malformed stub feature: final character {} not in expected set (+,-,?)".format(supported_type))
- supported_dict[key] = supported_type
- # Ensure we know the supported element
- if key not in self._KNOWN_QSUPPORTED_STUB_FEATURES:
- raise Exception(
- "unknown qSupported stub feature reported: %s" %
- key)
-
- return supported_dict
-
- def run_process_then_stop(self, run_seconds=1):
- # Tell the stub to continue.
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8"],
- True)
- context = self.expect_gdbremote_sequence()
-
- # Wait for run_seconds.
- time.sleep(run_seconds)
-
- # Send an interrupt, capture a T response.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: {}".format(chr(3)),
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]+)([^#]+)#[0-9a-fA-F]{2}$", "capture": {1: "stop_result"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- self.assertIsNotNone(context.get("stop_result"))
-
- return context
-
- def select_modifiable_register(self, reg_infos):
- """Find a register that can be read/written freely."""
- PREFERRED_REGISTER_NAMES = set(["rax", ])
-
- # First check for the first register from the preferred register name
- # set.
- alternative_register_index = None
-
- self.assertIsNotNone(reg_infos)
- for reg_info in reg_infos:
- if ("name" in reg_info) and (
- reg_info["name"] in PREFERRED_REGISTER_NAMES):
- # We found a preferred register. Use it.
- return reg_info["lldb_register_index"]
- if ("generic" in reg_info) and (reg_info["generic"] == "fp" or
- reg_info["generic"] == "arg1"):
- # A frame pointer or first arg register will do as a
- # register to modify temporarily.
- alternative_register_index = reg_info["lldb_register_index"]
-
- # We didn't find a preferred register. Return whatever alternative register
- # we found, if any.
- return alternative_register_index
-
- def extract_registers_from_stop_notification(self, stop_key_vals_text):
- self.assertIsNotNone(stop_key_vals_text)
- kv_dict = self.parse_key_val_dict(stop_key_vals_text)
-
- registers = {}
- for (key, val) in list(kv_dict.items()):
- if re.match(r"^[0-9a-fA-F]+$", key):
- registers[int(key, 16)] = val
- return registers
-
- def gather_register_infos(self):
- self.reset_test_sequence()
- self.add_register_info_collection_packets()
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.add_lldb_register_index(reg_infos)
-
- return reg_infos
-
- def find_generic_register_with_name(self, reg_infos, generic_name):
- self.assertIsNotNone(reg_infos)
- for reg_info in reg_infos:
- if ("generic" in reg_info) and (
- reg_info["generic"] == generic_name):
- return reg_info
- return None
-
- def decode_gdbremote_binary(self, encoded_bytes):
- decoded_bytes = ""
- i = 0
- while i < len(encoded_bytes):
- if encoded_bytes[i] == "}":
- # Handle escaped char.
- self.assertTrue(i + 1 < len(encoded_bytes))
- decoded_bytes += chr(ord(encoded_bytes[i + 1]) ^ 0x20)
- i += 2
- elif encoded_bytes[i] == "*":
- # Handle run length encoding.
- self.assertTrue(len(decoded_bytes) > 0)
- self.assertTrue(i + 1 < len(encoded_bytes))
- repeat_count = ord(encoded_bytes[i + 1]) - 29
- decoded_bytes += decoded_bytes[-1] * repeat_count
- i += 2
- else:
- decoded_bytes += encoded_bytes[i]
- i += 1
- return decoded_bytes
-
- def build_auxv_dict(self, endian, word_size, auxv_data):
- self.assertIsNotNone(endian)
- self.assertIsNotNone(word_size)
- self.assertIsNotNone(auxv_data)
-
- auxv_dict = {}
-
- # PowerPC64le's auxvec has a special key that must be ignored.
- # This special key may be used multiple times, resulting in
- # multiple key/value pairs with the same key, which would otherwise
- # break this test check for repeated keys.
- #
- # AT_IGNOREPPC = 22
- ignored_keys_for_arch = { 'powerpc64le' : [22] }
- arch = self.getArchitecture()
- ignore_keys = None
- if arch in ignored_keys_for_arch:
- ignore_keys = ignored_keys_for_arch[arch]
-
- while len(auxv_data) > 0:
- # Chop off key.
- raw_key = auxv_data[:word_size]
- auxv_data = auxv_data[word_size:]
-
- # Chop of value.
- raw_value = auxv_data[:word_size]
- auxv_data = auxv_data[word_size:]
-
- # Convert raw text from target endian.
- key = unpack_endian_binary_string(endian, raw_key)
- value = unpack_endian_binary_string(endian, raw_value)
-
- if ignore_keys and key in ignore_keys:
- continue
-
- # Handle ending entry.
- if key == 0:
- self.assertEqual(value, 0)
- return auxv_dict
-
- # The key should not already be present.
- self.assertFalse(key in auxv_dict)
- auxv_dict[key] = value
-
- self.fail(
- "should not reach here - implies required double zero entry not found")
- return auxv_dict
-
- def read_binary_data_in_chunks(self, command_prefix, chunk_length):
- """Collect command_prefix{offset:x},{chunk_length:x} until a single 'l' or 'l' with data is returned."""
- offset = 0
- done = False
- decoded_data = ""
-
- while not done:
- # Grab the next iteration of data.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- [
- "read packet: ${}{:x},{:x}:#00".format(
- command_prefix,
- offset,
- chunk_length),
- {
- "direction": "send",
- "regex": re.compile(
- r"^\$([^E])(.*)#[0-9a-fA-F]{2}$",
- re.MULTILINE | re.DOTALL),
- "capture": {
- 1: "response_type",
- 2: "content_raw"}}],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- response_type = context.get("response_type")
- self.assertIsNotNone(response_type)
- self.assertTrue(response_type in ["l", "m"])
-
- # Move offset along.
- offset += chunk_length
-
- # Figure out if we're done. We're done if the response type is l.
- done = response_type == "l"
-
- # Decode binary data.
- content_raw = context.get("content_raw")
- if content_raw and len(content_raw) > 0:
- self.assertIsNotNone(content_raw)
- decoded_data += self.decode_gdbremote_binary(content_raw)
- return decoded_data
-
- def add_interrupt_packets(self):
- self.test_sequence.add_log_lines([
- # Send the intterupt.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2})(.*)#[0-9a-fA-F]{2}$",
- "capture": {1: "stop_signo",
- 2: "stop_key_val_text"}},
- ], True)
-
- def parse_interrupt_packets(self, context):
- self.assertIsNotNone(context.get("stop_signo"))
- self.assertIsNotNone(context.get("stop_key_val_text"))
- return (int(context["stop_signo"], 16), self.parse_key_val_dict(
- context["stop_key_val_text"]))
-
- def add_QSaveRegisterState_packets(self, thread_id):
- if thread_id:
- # Use the thread suffix form.
- request = "read packet: $QSaveRegisterState;thread:{:x}#00".format(
- thread_id)
- else:
- request = "read packet: $QSaveRegisterState#00"
-
- self.test_sequence.add_log_lines([request,
- {"direction": "send",
- "regex": r"^\$(E?.*)#[0-9a-fA-F]{2}$",
- "capture": {1: "save_response"}},
- ],
- True)
-
- def parse_QSaveRegisterState_response(self, context):
- self.assertIsNotNone(context)
-
- save_response = context.get("save_response")
- self.assertIsNotNone(save_response)
-
- if len(save_response) < 1 or save_response[0] == "E":
- # error received
- return (False, None)
- else:
- return (True, int(save_response))
-
- def add_QRestoreRegisterState_packets(self, save_id, thread_id=None):
- if thread_id:
- # Use the thread suffix form.
- request = "read packet: $QRestoreRegisterState:{};thread:{:x}#00".format(
- save_id, thread_id)
- else:
- request = "read packet: $QRestoreRegisterState:{}#00".format(
- save_id)
-
- self.test_sequence.add_log_lines([
- request,
- "send packet: $OK#00"
- ], True)
-
- def flip_all_bits_in_each_register_value(
- self, reg_infos, endian, thread_id=None):
- self.assertIsNotNone(reg_infos)
-
- successful_writes = 0
- failed_writes = 0
-
- for reg_info in reg_infos:
- # Use the lldb register index added to the reg info. We're not necessarily
- # working off a full set of register infos, so an inferred register
- # index could be wrong.
- reg_index = reg_info["lldb_register_index"]
- self.assertIsNotNone(reg_index)
-
- reg_byte_size = int(reg_info["bitsize"]) // 8
- self.assertTrue(reg_byte_size > 0)
-
- # Handle thread suffix.
- if thread_id:
- p_request = "read packet: $p{:x};thread:{:x}#00".format(
- reg_index, thread_id)
- else:
- p_request = "read packet: $p{:x}#00".format(reg_index)
-
- # Read the existing value.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines([
- p_request,
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify the response length.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
- initial_reg_value = unpack_register_hex_unsigned(
- endian, p_response)
-
- # Flip the value by xoring with all 1s
- all_one_bits_raw = "ff" * (int(reg_info["bitsize"]) // 8)
- flipped_bits_int = initial_reg_value ^ int(all_one_bits_raw, 16)
- # print("reg (index={}, name={}): val={}, flipped bits (int={}, hex={:x})".format(reg_index, reg_info["name"], initial_reg_value, flipped_bits_int, flipped_bits_int))
-
- # Handle thread suffix for P.
- if thread_id:
- P_request = "read packet: $P{:x}={};thread:{:x}#00".format(
- reg_index, pack_register_hex(
- endian, flipped_bits_int, byte_size=reg_byte_size), thread_id)
- else:
- P_request = "read packet: $P{:x}={}#00".format(
- reg_index, pack_register_hex(
- endian, flipped_bits_int, byte_size=reg_byte_size))
-
- # Write the flipped value to the register.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines([P_request,
- {"direction": "send",
- "regex": r"^\$(OK|E[0-9a-fA-F]+)#[0-9a-fA-F]{2}",
- "capture": {1: "P_response"}},
- ],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Determine if the write succeeded. There are a handful of registers that can fail, or partially fail
- # (e.g. flags, segment selectors, etc.) due to register value restrictions. Don't worry about them
- # all flipping perfectly.
- P_response = context.get("P_response")
- self.assertIsNotNone(P_response)
- if P_response == "OK":
- successful_writes += 1
- else:
- failed_writes += 1
- # print("reg (index={}, name={}) write FAILED (error: {})".format(reg_index, reg_info["name"], P_response))
-
- # Read back the register value, ensure it matches the flipped
- # value.
- if P_response == "OK":
- self.reset_test_sequence()
- self.test_sequence.add_log_lines([
- p_request,
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- verify_p_response_raw = context.get("p_response")
- self.assertIsNotNone(verify_p_response_raw)
- verify_bits = unpack_register_hex_unsigned(
- endian, verify_p_response_raw)
-
- if verify_bits != flipped_bits_int:
- # Some registers, like mxcsrmask and others, will permute what's written. Adjust succeed/fail counts.
- # print("reg (index={}, name={}): read verify FAILED: wrote {:x}, verify read back {:x}".format(reg_index, reg_info["name"], flipped_bits_int, verify_bits))
- successful_writes -= 1
- failed_writes += 1
-
- return (successful_writes, failed_writes)
-
- def is_bit_flippable_register(self, reg_info):
- if not reg_info:
- return False
- if not "set" in reg_info:
- return False
- if reg_info["set"] != "General Purpose Registers":
- return False
- if ("container-regs" in reg_info) and (
- len(reg_info["container-regs"]) > 0):
- # Don't try to bit flip registers contained in another register.
- return False
- if re.match("^.s$", reg_info["name"]):
- # This is a 2-letter register name that ends in "s", like a segment register.
- # Don't try to bit flip these.
- return False
- if re.match("^(c|)psr$", reg_info["name"]):
- # This is an ARM program status register; don't flip it.
- return False
- # Okay, this looks fine-enough.
- return True
-
- def read_register_values(self, reg_infos, endian, thread_id=None):
- self.assertIsNotNone(reg_infos)
- values = {}
-
- for reg_info in reg_infos:
- # We append a register index when load reg infos so we can work
- # with subsets.
- reg_index = reg_info.get("lldb_register_index")
- self.assertIsNotNone(reg_index)
-
- # Handle thread suffix.
- if thread_id:
- p_request = "read packet: $p{:x};thread:{:x}#00".format(
- reg_index, thread_id)
- else:
- p_request = "read packet: $p{:x}#00".format(reg_index)
-
- # Read it with p.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines([
- p_request,
- {"direction": "send", "regex": r"^\$([0-9a-fA-F]+)#", "capture": {1: "p_response"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Convert value from target endian to integral.
- p_response = context.get("p_response")
- self.assertIsNotNone(p_response)
- self.assertTrue(len(p_response) > 0)
- self.assertFalse(p_response[0] == "E")
-
- values[reg_index] = unpack_register_hex_unsigned(
- endian, p_response)
-
- return values
-
- def add_vCont_query_packets(self):
- self.test_sequence.add_log_lines(["read packet: $vCont?#49",
- {"direction": "send",
- "regex": r"^\$(vCont)?(.*)#[0-9a-fA-F]{2}$",
- "capture": {2: "vCont_query_response"}},
- ],
- True)
-
- def parse_vCont_query_response(self, context):
- self.assertIsNotNone(context)
- vCont_query_response = context.get("vCont_query_response")
-
- # Handle case of no vCont support at all - in which case the capture
- # group will be none or zero length.
- if not vCont_query_response or len(vCont_query_response) == 0:
- return {}
-
- return {key: 1 for key in vCont_query_response.split(
- ";") if key and len(key) > 0}
-
- def count_single_steps_until_true(
- self,
- thread_id,
- predicate,
- args,
- max_step_count=100,
- use_Hc_packet=True,
- step_instruction="s"):
- """Used by single step test that appears in a few different contexts."""
- single_step_count = 0
-
- while single_step_count < max_step_count:
- self.assertIsNotNone(thread_id)
-
- # Build the packet for the single step instruction. We replace
- # {thread}, if present, with the thread_id.
- step_packet = "read packet: ${}#00".format(
- re.sub(r"{thread}", "{:x}".format(thread_id), step_instruction))
- # print("\nstep_packet created: {}\n".format(step_packet))
-
- # Single step.
- self.reset_test_sequence()
- if use_Hc_packet:
- self.test_sequence.add_log_lines(
- [ # Set the continue thread.
- "read packet: $Hc{0:x}#00".format(thread_id),
- "send packet: $OK#00",
- ], True)
- self.test_sequence.add_log_lines([
- # Single step.
- step_packet,
- # "read packet: $vCont;s:{0:x}#00".format(thread_id),
- # Expect a breakpoint stop report.
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);",
- "capture": {1: "stop_signo",
- 2: "stop_thread_id"}},
- ], True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- self.assertIsNotNone(context.get("stop_signo"))
- self.assertEqual(int(context.get("stop_signo"), 16),
- lldbutil.get_signal_number('SIGTRAP'))
-
- single_step_count += 1
-
- # See if the predicate is true. If so, we're done.
- if predicate(args):
- return (True, single_step_count)
-
- # The predicate didn't return true within the runaway step count.
- return (False, single_step_count)
-
- def g_c1_c2_contents_are(self, args):
- """Used by single step test that appears in a few different contexts."""
- g_c1_address = args["g_c1_address"]
- g_c2_address = args["g_c2_address"]
- expected_g_c1 = args["expected_g_c1"]
- expected_g_c2 = args["expected_g_c2"]
-
- # Read g_c1 and g_c2 contents.
- self.reset_test_sequence()
- self.test_sequence.add_log_lines(
- ["read packet: $m{0:x},{1:x}#00".format(g_c1_address, 1),
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$", "capture": {1: "g_c1_contents"}},
- "read packet: $m{0:x},{1:x}#00".format(g_c2_address, 1),
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$", "capture": {1: "g_c2_contents"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Check if what we read from inferior memory is what we are expecting.
- self.assertIsNotNone(context.get("g_c1_contents"))
- self.assertIsNotNone(context.get("g_c2_contents"))
-
- return (seven.unhexlify(context.get("g_c1_contents")) == expected_g_c1) and (
- seven.unhexlify(context.get("g_c2_contents")) == expected_g_c2)
-
- def single_step_only_steps_one_instruction(
- self, use_Hc_packet=True, step_instruction="s"):
- """Used by single step test that appears in a few different contexts."""
- # Start up the inferior.
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=[
- "get-code-address-hex:swap_chars",
- "get-data-address-hex:g_c1",
- "get-data-address-hex:g_c2",
- "sleep:1",
- "call-function:swap_chars",
- "sleep:5"])
-
- # Run the process
- self.test_sequence.add_log_lines(
- [ # Start running after initial stop.
- "read packet: $c#63",
- # Match output line that prints the memory address of the function call entry point.
- # Note we require launch-only testing so we can get inferior otuput.
- {"type": "output_match", "regex": r"^code address: 0x([0-9a-fA-F]+)\r\ndata address: 0x([0-9a-fA-F]+)\r\ndata address: 0x([0-9a-fA-F]+)\r\n$",
- "capture": {1: "function_address", 2: "g_c1_address", 3: "g_c2_address"}},
- # Now stop the inferior.
- "read packet: {}".format(chr(3)),
- # And wait for the stop notification.
- {"direction": "send", "regex": r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+);", "capture": {1: "stop_signo", 2: "stop_thread_id"}}],
- True)
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Grab the main thread id.
- self.assertIsNotNone(context.get("stop_thread_id"))
- main_thread_id = int(context.get("stop_thread_id"), 16)
-
- # Grab the function address.
- self.assertIsNotNone(context.get("function_address"))
- function_address = int(context.get("function_address"), 16)
-
- # Grab the data addresses.
- self.assertIsNotNone(context.get("g_c1_address"))
- g_c1_address = int(context.get("g_c1_address"), 16)
-
- self.assertIsNotNone(context.get("g_c2_address"))
- g_c2_address = int(context.get("g_c2_address"), 16)
-
- # Set a breakpoint at the given address.
- if self.getArchitecture() == "arm":
- # TODO: Handle case when setting breakpoint in thumb code
- BREAKPOINT_KIND = 4
- else:
- BREAKPOINT_KIND = 1
- self.reset_test_sequence()
- self.add_set_breakpoint_packets(
- function_address,
- do_continue=True,
- breakpoint_kind=BREAKPOINT_KIND)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Remove the breakpoint.
- self.reset_test_sequence()
- self.add_remove_breakpoint_packets(
- function_address, breakpoint_kind=BREAKPOINT_KIND)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Verify g_c1 and g_c2 match expected initial state.
- args = {}
- args["g_c1_address"] = g_c1_address
- args["g_c2_address"] = g_c2_address
- args["expected_g_c1"] = "0"
- args["expected_g_c2"] = "1"
-
- self.assertTrue(self.g_c1_c2_contents_are(args))
-
- # Verify we take only a small number of steps to hit the first state.
- # Might need to work through function entry prologue code.
- args["expected_g_c1"] = "1"
- args["expected_g_c2"] = "1"
- (state_reached,
- step_count) = self.count_single_steps_until_true(main_thread_id,
- self.g_c1_c2_contents_are,
- args,
- max_step_count=25,
- use_Hc_packet=use_Hc_packet,
- step_instruction=step_instruction)
- self.assertTrue(state_reached)
-
- # Verify we hit the next state.
- args["expected_g_c1"] = "1"
- args["expected_g_c2"] = "0"
- (state_reached,
- step_count) = self.count_single_steps_until_true(main_thread_id,
- self.g_c1_c2_contents_are,
- args,
- max_step_count=5,
- use_Hc_packet=use_Hc_packet,
- step_instruction=step_instruction)
- self.assertTrue(state_reached)
- expected_step_count = 1
- arch = self.getArchitecture()
-
- # MIPS required "3" (ADDIU, SB, LD) machine instructions for updation
- # of variable value
- if re.match("mips", arch):
- expected_step_count = 3
- # S390X requires "2" (LARL, MVI) machine instructions for updation of
- # variable value
- if re.match("s390x", arch):
- expected_step_count = 2
- self.assertEqual(step_count, expected_step_count)
-
- # Verify we hit the next state.
- args["expected_g_c1"] = "0"
- args["expected_g_c2"] = "0"
- (state_reached,
- step_count) = self.count_single_steps_until_true(main_thread_id,
- self.g_c1_c2_contents_are,
- args,
- max_step_count=5,
- use_Hc_packet=use_Hc_packet,
- step_instruction=step_instruction)
- self.assertTrue(state_reached)
- self.assertEqual(step_count, expected_step_count)
-
- # Verify we hit the next state.
- args["expected_g_c1"] = "0"
- args["expected_g_c2"] = "1"
- (state_reached,
- step_count) = self.count_single_steps_until_true(main_thread_id,
- self.g_c1_c2_contents_are,
- args,
- max_step_count=5,
- use_Hc_packet=use_Hc_packet,
- step_instruction=step_instruction)
- self.assertTrue(state_reached)
- self.assertEqual(step_count, expected_step_count)
-
- def maybe_strict_output_regex(self, regex):
- return '.*' + regex + \
- '.*' if lldbplatformutil.hasChattyStderr(self) else '^' + regex + '$'
-
- def install_and_create_launch_args(self):
- exe_path = self.getBuildArtifact("a.out")
- if not lldb.remote_platform:
- return [exe_path]
- remote_path = lldbutil.append_to_process_working_directory(self,
- os.path.basename(exe_path))
- remote_file_spec = lldb.SBFileSpec(remote_path, False)
- err = lldb.remote_platform.Install(lldb.SBFileSpec(exe_path, True),
- remote_file_spec)
- if err.Fail():
- raise Exception("remote_platform.Install('%s', '%s') failed: %s" %
- (exe_path, remote_path, err))
- return [remote_path]
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/Makefile
deleted file mode 100644
index a47e2797fd8..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/Makefile
+++ /dev/null
@@ -1,8 +0,0 @@
-LEVEL = ../../../make
-
-CFLAGS_EXTRAS += -D__STDC_LIMIT_MACROS -D__STDC_FORMAT_MACROS -std=c++11
-# LD_EXTRAS := -lpthread
-CXX_SOURCES := main.cpp
-MAKE_DSYM :=NO
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteAbort.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteAbort.py
deleted file mode 100644
index e905a85cc4d..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteAbort.py
+++ /dev/null
@@ -1,46 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import signal
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteAbort(gdbremote_testcase.GdbRemoteTestCaseBase):
- mydir = TestBase.compute_mydir(__file__)
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def inferior_abort_received(self):
- procs = self.prep_debug_monitor_and_inferior(inferior_args=["abort"])
- self.assertIsNotNone(procs)
-
- self.test_sequence.add_log_lines(["read packet: $vCont;c#a8",
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2}).*#[0-9a-fA-F]{2}$",
- "capture": {1: "hex_exit_code"}},
- ],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- hex_exit_code = context.get("hex_exit_code")
- self.assertIsNotNone(hex_exit_code)
- self.assertEqual(int(hex_exit_code, 16),
- lldbutil.get_signal_number('SIGABRT'))
-
- @debugserver_test
- def test_inferior_abort_received_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.inferior_abort_received()
-
- @llgs_test
- # std::abort() on <= API 16 raises SIGSEGV - b.android.com/179836
- @expectedFailureAndroid(api_levels=list(range(16 + 1)))
- def test_inferior_abort_received_llgs(self):
- self.init_llgs_test()
- self.build()
- self.inferior_abort_received()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteSegFault.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteSegFault.py
deleted file mode 100644
index 316d5d7b82b..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/TestGdbRemoteSegFault.py
+++ /dev/null
@@ -1,46 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import signal
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteSegFault(gdbremote_testcase.GdbRemoteTestCaseBase):
- mydir = TestBase.compute_mydir(__file__)
-
- GDB_REMOTE_STOP_CODE_BAD_ACCESS = 0x91
-
- @skipIfDarwinEmbedded # <rdar://problem/34539270> lldb-server tests not updated to work on ios etc yet
- def inferior_seg_fault_received(self, expected_signo):
- procs = self.prep_debug_monitor_and_inferior(
- inferior_args=["segfault"])
- self.assertIsNotNone(procs)
-
- self.test_sequence.add_log_lines(["read packet: $vCont;c#a8",
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2}).*#[0-9a-fA-F]{2}$",
- "capture": {1: "hex_exit_code"}},
- ],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- hex_exit_code = context.get("hex_exit_code")
- self.assertIsNotNone(hex_exit_code)
- self.assertEqual(int(hex_exit_code, 16), expected_signo)
-
- @debugserver_test
- def test_inferior_seg_fault_received_debugserver(self):
- self.init_debugserver_test()
- self.build()
- self.inferior_seg_fault_received(self.GDB_REMOTE_STOP_CODE_BAD_ACCESS)
-
- @llgs_test
- def test_inferior_seg_fault_received_llgs(self):
- self.init_llgs_test()
- self.build()
- self.inferior_seg_fault_received(lldbutil.get_signal_number('SIGSEGV'))
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/main.cpp
deleted file mode 100644
index ced7f712508..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/inferior-crash/main.cpp
+++ /dev/null
@@ -1,31 +0,0 @@
-#include <cstdlib>
-#include <cstring>
-#include <iostream>
-
-namespace {
-const char *const SEGFAULT_COMMAND = "segfault";
-const char *const ABORT_COMMAND = "abort";
-}
-
-int main(int argc, char **argv) {
- if (argc < 2) {
- std::cout << "expected at least one command provided on the command line"
- << std::endl;
- }
-
- // Process command line args.
- for (int i = 1; i < argc; ++i) {
- const char *const command = argv[i];
- if (std::strstr(command, SEGFAULT_COMMAND)) {
- // Perform a null pointer access.
- int *const null_int_ptr = nullptr;
- *null_int_ptr = 0xDEAD;
- } else if (std::strstr(command, ABORT_COMMAND)) {
- std::abort();
- } else {
- std::cout << "Unsupported command: " << command << std::endl;
- }
- }
-
- return 0;
-}
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/lldbgdbserverutils.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/lldbgdbserverutils.py
deleted file mode 100644
index 8dd146ae152..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/lldbgdbserverutils.py
+++ /dev/null
@@ -1,945 +0,0 @@
-"""Module for supporting unit testing of the lldb-server debug monitor exe.
-"""
-
-from __future__ import division, print_function
-
-
-import os
-import os.path
-import platform
-import re
-import six
-import socket_packet_pump
-import subprocess
-import time
-from lldbsuite.test.lldbtest import *
-
-from six.moves import queue
-
-
-def _get_debug_monitor_from_lldb(lldb_exe, debug_monitor_basename):
- """Return the debug monitor exe path given the lldb exe path.
-
- This method attempts to construct a valid debug monitor exe name
- from a given lldb exe name. It will return None if the synthesized
- debug monitor name is not found to exist.
-
- The debug monitor exe path is synthesized by taking the directory
- of the lldb exe, and replacing the portion of the base name that
- matches "lldb" (case insensitive) and replacing with the value of
- debug_monitor_basename.
-
- Args:
- lldb_exe: the path to an lldb executable.
-
- debug_monitor_basename: the base name portion of the debug monitor
- that will replace 'lldb'.
-
- Returns:
- A path to the debug monitor exe if it is found to exist; otherwise,
- returns None.
-
- """
- if not lldb_exe:
- return None
-
- exe_dir = os.path.dirname(lldb_exe)
- exe_base = os.path.basename(lldb_exe)
-
- # we'll rebuild the filename by replacing lldb with
- # the debug monitor basename, keeping any prefix or suffix in place.
- regex = re.compile(r"lldb", re.IGNORECASE)
- new_base = regex.sub(debug_monitor_basename, exe_base)
-
- debug_monitor_exe = os.path.join(exe_dir, new_base)
- if os.path.exists(debug_monitor_exe):
- return debug_monitor_exe
-
- new_base = regex.sub(
- 'LLDB.framework/Versions/A/Resources/' +
- debug_monitor_basename,
- exe_base)
- debug_monitor_exe = os.path.join(exe_dir, new_base)
- if os.path.exists(debug_monitor_exe):
- return debug_monitor_exe
-
- return None
-
-
-def get_lldb_server_exe():
- """Return the lldb-server exe path.
-
- Returns:
- A path to the lldb-server exe if it is found to exist; otherwise,
- returns None.
- """
- if "LLDB_DEBUGSERVER_PATH" in os.environ:
- return os.environ["LLDB_DEBUGSERVER_PATH"]
-
- return _get_debug_monitor_from_lldb(
- lldbtest_config.lldbExec, "lldb-server")
-
-
-def get_debugserver_exe():
- """Return the debugserver exe path.
-
- Returns:
- A path to the debugserver exe if it is found to exist; otherwise,
- returns None.
- """
- if "LLDB_DEBUGSERVER_PATH" in os.environ:
- return os.environ["LLDB_DEBUGSERVER_PATH"]
-
- return _get_debug_monitor_from_lldb(
- lldbtest_config.lldbExec, "debugserver")
-
-_LOG_LINE_REGEX = re.compile(r'^(lldb-server|debugserver)\s+<\s*(\d+)>' +
- '\s+(read|send)\s+packet:\s+(.+)$')
-
-
-def _is_packet_lldb_gdbserver_input(packet_type, llgs_input_is_read):
- """Return whether a given packet is input for lldb-gdbserver.
-
- Args:
- packet_type: a string indicating 'send' or 'receive', from a
- gdbremote packet protocol log.
-
- llgs_input_is_read: true if lldb-gdbserver input (content sent to
- lldb-gdbserver) is listed as 'read' or 'send' in the packet
- log entry.
-
- Returns:
- True if the packet should be considered input for lldb-gdbserver; False
- otherwise.
- """
- if packet_type == 'read':
- # when llgs is the read side, then a read packet is meant for
- # input to llgs (when captured from the llgs/debugserver exe).
- return llgs_input_is_read
- elif packet_type == 'send':
- # when llgs is the send side, then a send packet is meant to
- # be input to llgs (when captured from the lldb exe).
- return not llgs_input_is_read
- else:
- # don't understand what type of packet this is
- raise "Unknown packet type: {}".format(packet_type)
-
-
-def handle_O_packet(context, packet_contents, logger):
- """Handle O packets."""
- if (not packet_contents) or (len(packet_contents) < 1):
- return False
- elif packet_contents[0] != "O":
- return False
- elif packet_contents == "OK":
- return False
-
- new_text = gdbremote_hex_decode_string(packet_contents[1:])
- context["O_content"] += new_text
- context["O_count"] += 1
-
- if logger:
- logger.debug(
- "text: new \"{}\", cumulative: \"{}\"".format(
- new_text, context["O_content"]))
-
- return True
-
-_STRIP_CHECKSUM_REGEX = re.compile(r'#[0-9a-fA-F]{2}$')
-_STRIP_COMMAND_PREFIX_REGEX = re.compile(r"^\$")
-_STRIP_COMMAND_PREFIX_M_REGEX = re.compile(r"^\$m")
-
-
-def assert_packets_equal(asserter, actual_packet, expected_packet):
- # strip off the checksum digits of the packet. When we're in
- # no-ack mode, the # checksum is ignored, and should not be cause
- # for a mismatched packet.
- actual_stripped = _STRIP_CHECKSUM_REGEX.sub('', actual_packet)
- expected_stripped = _STRIP_CHECKSUM_REGEX.sub('', expected_packet)
- asserter.assertEqual(actual_stripped, expected_stripped)
-
-
-def expect_lldb_gdbserver_replay(
- asserter,
- sock,
- test_sequence,
- pump_queues,
- timeout_seconds,
- logger=None):
- """Replay socket communication with lldb-gdbserver and verify responses.
-
- Args:
- asserter: the object providing assertEqual(first, second, msg=None), e.g. TestCase instance.
-
- sock: the TCP socket connected to the lldb-gdbserver exe.
-
- test_sequence: a GdbRemoteTestSequence instance that describes
- the messages sent to the gdb remote and the responses
- expected from it.
-
- timeout_seconds: any response taking more than this number of
- seconds will cause an exception to be raised.
-
- logger: a Python logger instance.
-
- Returns:
- The context dictionary from running the given gdbremote
- protocol sequence. This will contain any of the capture
- elements specified to any GdbRemoteEntry instances in
- test_sequence.
-
- The context will also contain an entry, context["O_content"]
- which contains the text from the inferior received via $O
- packets. $O packets should not attempt to be matched
- directly since they are not entirely deterministic as to
- how many arrive and how much text is in each one.
-
- context["O_count"] will contain an integer of the number of
- O packets received.
- """
-
- # Ensure we have some work to do.
- if len(test_sequence.entries) < 1:
- return {}
-
- context = {"O_count": 0, "O_content": ""}
- with socket_packet_pump.SocketPacketPump(sock, pump_queues, logger) as pump:
- # Grab the first sequence entry.
- sequence_entry = test_sequence.entries.pop(0)
-
- # While we have an active sequence entry, send messages
- # destined for the stub and collect/match/process responses
- # expected from the stub.
- while sequence_entry:
- if sequence_entry.is_send_to_remote():
- # This is an entry to send to the remote debug monitor.
- send_packet = sequence_entry.get_send_packet()
- if logger:
- if len(send_packet) == 1 and send_packet[0] == chr(3):
- packet_desc = "^C"
- else:
- packet_desc = send_packet
- logger.info(
- "sending packet to remote: {}".format(packet_desc))
- sock.sendall(send_packet.encode())
- else:
- # This is an entry expecting to receive content from the remote
- # debug monitor.
-
- # We'll pull from (and wait on) the queue appropriate for the type of matcher.
- # We keep separate queues for process output (coming from non-deterministic
- # $O packet division) and for all other packets.
- if sequence_entry.is_output_matcher():
- try:
- # Grab next entry from the output queue.
- content = pump_queues.output_queue().get(True, timeout_seconds)
- except queue.Empty:
- if logger:
- logger.warning(
- "timeout waiting for stub output (accumulated output:{})".format(
- pump.get_accumulated_output()))
- raise Exception(
- "timed out while waiting for output match (accumulated output: {})".format(
- pump.get_accumulated_output()))
- else:
- try:
- content = pump_queues.packet_queue().get(True, timeout_seconds)
- except queue.Empty:
- if logger:
- logger.warning(
- "timeout waiting for packet match (receive buffer: {})".format(
- pump.get_receive_buffer()))
- raise Exception(
- "timed out while waiting for packet match (receive buffer: {})".format(
- pump.get_receive_buffer()))
-
- # Give the sequence entry the opportunity to match the content.
- # Output matchers might match or pass after more output accumulates.
- # Other packet types generally must match.
- asserter.assertIsNotNone(content)
- context = sequence_entry.assert_match(
- asserter, content, context=context)
-
- # Move on to next sequence entry as needed. Some sequence entries support executing multiple
- # times in different states (for looping over query/response
- # packets).
- if sequence_entry.is_consumed():
- if len(test_sequence.entries) > 0:
- sequence_entry = test_sequence.entries.pop(0)
- else:
- sequence_entry = None
-
- # Fill in the O_content entries.
- context["O_count"] = 1
- context["O_content"] = pump.get_accumulated_output()
-
- return context
-
-
-def gdbremote_hex_encode_string(str):
- output = ''
- for c in str:
- output += '{0:02x}'.format(ord(c))
- return output
-
-
-def gdbremote_hex_decode_string(str):
- return str.decode("hex")
-
-
-def gdbremote_packet_encode_string(str):
- checksum = 0
- for c in str:
- checksum += ord(c)
- return '$' + str + '#{0:02x}'.format(checksum % 256)
-
-
-def build_gdbremote_A_packet(args_list):
- """Given a list of args, create a properly-formed $A packet containing each arg.
- """
- payload = "A"
-
- # build the arg content
- arg_index = 0
- for arg in args_list:
- # Comma-separate the args.
- if arg_index > 0:
- payload += ','
-
- # Hex-encode the arg.
- hex_arg = gdbremote_hex_encode_string(arg)
-
- # Build the A entry.
- payload += "{},{},{}".format(len(hex_arg), arg_index, hex_arg)
-
- # Next arg index, please.
- arg_index += 1
-
- # return the packetized payload
- return gdbremote_packet_encode_string(payload)
-
-
-def parse_reg_info_response(response_packet):
- if not response_packet:
- raise Exception("response_packet cannot be None")
-
- # Strip off prefix $ and suffix #xx if present.
- response_packet = _STRIP_COMMAND_PREFIX_REGEX.sub("", response_packet)
- response_packet = _STRIP_CHECKSUM_REGEX.sub("", response_packet)
-
- # Build keyval pairs
- values = {}
- for kv in response_packet.split(";"):
- if len(kv) < 1:
- continue
- (key, val) = kv.split(':')
- values[key] = val
-
- return values
-
-
-def parse_threadinfo_response(response_packet):
- if not response_packet:
- raise Exception("response_packet cannot be None")
-
- # Strip off prefix $ and suffix #xx if present.
- response_packet = _STRIP_COMMAND_PREFIX_M_REGEX.sub("", response_packet)
- response_packet = _STRIP_CHECKSUM_REGEX.sub("", response_packet)
-
- # Return list of thread ids
- return [int(thread_id_hex, 16) for thread_id_hex in response_packet.split(
- ",") if len(thread_id_hex) > 0]
-
-
-def unpack_endian_binary_string(endian, value_string):
- """Unpack a gdb-remote binary (post-unescaped, i.e. not escaped) response to an unsigned int given endianness of the inferior."""
- if not endian:
- raise Exception("endian cannot be None")
- if not value_string or len(value_string) < 1:
- raise Exception("value_string cannot be None or empty")
-
- if endian == 'little':
- value = 0
- i = 0
- while len(value_string) > 0:
- value += (ord(value_string[0]) << i)
- value_string = value_string[1:]
- i += 8
- return value
- elif endian == 'big':
- value = 0
- while len(value_string) > 0:
- value = (value << 8) + ord(value_string[0])
- value_string = value_string[1:]
- return value
- else:
- # pdp is valid but need to add parse code once needed.
- raise Exception("unsupported endian:{}".format(endian))
-
-
-def unpack_register_hex_unsigned(endian, value_string):
- """Unpack a gdb-remote $p-style response to an unsigned int given endianness of inferior."""
- if not endian:
- raise Exception("endian cannot be None")
- if not value_string or len(value_string) < 1:
- raise Exception("value_string cannot be None or empty")
-
- if endian == 'little':
- value = 0
- i = 0
- while len(value_string) > 0:
- value += (int(value_string[0:2], 16) << i)
- value_string = value_string[2:]
- i += 8
- return value
- elif endian == 'big':
- return int(value_string, 16)
- else:
- # pdp is valid but need to add parse code once needed.
- raise Exception("unsupported endian:{}".format(endian))
-
-
-def pack_register_hex(endian, value, byte_size=None):
- """Unpack a gdb-remote $p-style response to an unsigned int given endianness of inferior."""
- if not endian:
- raise Exception("endian cannot be None")
-
- if endian == 'little':
- # Create the litt-endian return value.
- retval = ""
- while value != 0:
- retval = retval + "{:02x}".format(value & 0xff)
- value = value >> 8
- if byte_size:
- # Add zero-fill to the right/end (MSB side) of the value.
- retval += "00" * (byte_size - len(retval) // 2)
- return retval
-
- elif endian == 'big':
- retval = ""
- while value != 0:
- retval = "{:02x}".format(value & 0xff) + retval
- value = value >> 8
- if byte_size:
- # Add zero-fill to the left/front (MSB side) of the value.
- retval = ("00" * (byte_size - len(retval) // 2)) + retval
- return retval
-
- else:
- # pdp is valid but need to add parse code once needed.
- raise Exception("unsupported endian:{}".format(endian))
-
-
-class GdbRemoteEntryBase(object):
-
- def is_output_matcher(self):
- return False
-
-
-class GdbRemoteEntry(GdbRemoteEntryBase):
-
- def __init__(
- self,
- is_send_to_remote=True,
- exact_payload=None,
- regex=None,
- capture=None,
- expect_captures=None):
- """Create an entry representing one piece of the I/O to/from a gdb remote debug monitor.
-
- Args:
-
- is_send_to_remote: True if this entry is a message to be
- sent to the gdbremote debug monitor; False if this
- entry represents text to be matched against the reply
- from the gdbremote debug monitor.
-
- exact_payload: if not None, then this packet is an exact
- send (when sending to the remote) or an exact match of
- the response from the gdbremote. The checksums are
- ignored on exact match requests since negotiation of
- no-ack makes the checksum content essentially
- undefined.
-
- regex: currently only valid for receives from gdbremote.
- When specified (and only if exact_payload is None),
- indicates the gdbremote response must match the given
- regex. Match groups in the regex can be used for two
- different purposes: saving the match (see capture
- arg), or validating that a match group matches a
- previously established value (see expect_captures). It
- is perfectly valid to have just a regex arg and to
- specify neither capture or expect_captures args. This
- arg only makes sense if exact_payload is not
- specified.
-
- capture: if specified, is a dictionary of regex match
- group indices (should start with 1) to variable names
- that will store the capture group indicated by the
- index. For example, {1:"thread_id"} will store capture
- group 1's content in the context dictionary where
- "thread_id" is the key and the match group value is
- the value. The value stored off can be used later in a
- expect_captures expression. This arg only makes sense
- when regex is specified.
-
- expect_captures: if specified, is a dictionary of regex
- match group indices (should start with 1) to variable
- names, where the match group should match the value
- existing in the context at the given variable name.
- For example, {2:"thread_id"} indicates that the second
- match group must match the value stored under the
- context's previously stored "thread_id" key. This arg
- only makes sense when regex is specified.
- """
- self._is_send_to_remote = is_send_to_remote
- self.exact_payload = exact_payload
- self.regex = regex
- self.capture = capture
- self.expect_captures = expect_captures
-
- def is_send_to_remote(self):
- return self._is_send_to_remote
-
- def is_consumed(self):
- # For now, all packets are consumed after first use.
- return True
-
- def get_send_packet(self):
- if not self.is_send_to_remote():
- raise Exception(
- "get_send_packet() called on GdbRemoteEntry that is not a send-to-remote packet")
- if not self.exact_payload:
- raise Exception(
- "get_send_packet() called on GdbRemoteEntry but it doesn't have an exact payload")
- return self.exact_payload
-
- def _assert_exact_payload_match(self, asserter, actual_packet):
- assert_packets_equal(asserter, actual_packet, self.exact_payload)
- return None
-
- def _assert_regex_match(self, asserter, actual_packet, context):
- # Ensure the actual packet matches from the start of the actual packet.
- match = self.regex.match(actual_packet)
- if not match:
- asserter.fail(
- "regex '{}' failed to match against content '{}'".format(
- self.regex.pattern, actual_packet))
-
- if self.capture:
- # Handle captures.
- for group_index, var_name in list(self.capture.items()):
- capture_text = match.group(group_index)
- # It is okay for capture text to be None - which it will be if it is a group that can match nothing.
- # The user must be okay with it since the regex itself matched
- # above.
- context[var_name] = capture_text
-
- if self.expect_captures:
- # Handle comparing matched groups to context dictionary entries.
- for group_index, var_name in list(self.expect_captures.items()):
- capture_text = match.group(group_index)
- if not capture_text:
- raise Exception(
- "No content to expect for group index {}".format(group_index))
- asserter.assertEqual(capture_text, context[var_name])
-
- return context
-
- def assert_match(self, asserter, actual_packet, context=None):
- # This only makes sense for matching lines coming from the
- # remote debug monitor.
- if self.is_send_to_remote():
- raise Exception(
- "Attempted to match a packet being sent to the remote debug monitor, doesn't make sense.")
-
- # Create a new context if needed.
- if not context:
- context = {}
-
- # If this is an exact payload, ensure they match exactly,
- # ignoring the packet checksum which is optional for no-ack
- # mode.
- if self.exact_payload:
- self._assert_exact_payload_match(asserter, actual_packet)
- return context
- elif self.regex:
- return self._assert_regex_match(asserter, actual_packet, context)
- else:
- raise Exception(
- "Don't know how to match a remote-sent packet when exact_payload isn't specified.")
-
-
-class MultiResponseGdbRemoteEntry(GdbRemoteEntryBase):
- """Represents a query/response style packet.
-
- Assumes the first item is sent to the gdb remote.
- An end sequence regex indicates the end of the query/response
- packet sequence. All responses up through (but not including) the
- end response are stored in a context variable.
-
- Settings accepted from params:
-
- next_query or query: required. The typical query packet without the $ prefix or #xx suffix.
- If there is a special first packet to start the iteration query, see the
- first_query key.
-
- first_query: optional. If the first query requires a special query command, specify
- it with this key. Do not specify the $ prefix or #xx suffix.
-
- append_iteration_suffix: defaults to False. Specify True if the 0-based iteration
- index should be appended as a suffix to the command. e.g. qRegisterInfo with
- this key set true will generate query packets of qRegisterInfo0, qRegisterInfo1,
- etc.
-
- end_regex: required. Specifies a compiled regex object that will match the full text
- of any response that signals an end to the iteration. It must include the
- initial $ and ending #xx and must match the whole packet.
-
- save_key: required. Specifies the key within the context where an array will be stored.
- Each packet received from the gdb remote that does not match the end_regex will get
- appended to the array stored within the context at that key.
-
- runaway_response_count: optional. Defaults to 10000. If this many responses are retrieved,
- assume there is something wrong with either the response collection or the ending
- detection regex and throw an exception.
- """
-
- def __init__(self, params):
- self._next_query = params.get("next_query", params.get("query"))
- if not self._next_query:
- raise "either next_query or query key must be specified for MultiResponseGdbRemoteEntry"
-
- self._first_query = params.get("first_query", self._next_query)
- self._append_iteration_suffix = params.get(
- "append_iteration_suffix", False)
- self._iteration = 0
- self._end_regex = params["end_regex"]
- self._save_key = params["save_key"]
- self._runaway_response_count = params.get(
- "runaway_response_count", 10000)
- self._is_send_to_remote = True
- self._end_matched = False
-
- def is_send_to_remote(self):
- return self._is_send_to_remote
-
- def get_send_packet(self):
- if not self.is_send_to_remote():
- raise Exception(
- "get_send_packet() called on MultiResponseGdbRemoteEntry that is not in the send state")
- if self._end_matched:
- raise Exception(
- "get_send_packet() called on MultiResponseGdbRemoteEntry but end of query/response sequence has already been seen.")
-
- # Choose the first or next query for the base payload.
- if self._iteration == 0 and self._first_query:
- payload = self._first_query
- else:
- payload = self._next_query
-
- # Append the suffix as needed.
- if self._append_iteration_suffix:
- payload += "%x" % self._iteration
-
- # Keep track of the iteration.
- self._iteration += 1
-
- # Now that we've given the query packet, flip the mode to
- # receive/match.
- self._is_send_to_remote = False
-
- # Return the result, converted to packet form.
- return gdbremote_packet_encode_string(payload)
-
- def is_consumed(self):
- return self._end_matched
-
- def assert_match(self, asserter, actual_packet, context=None):
- # This only makes sense for matching lines coming from the remote debug
- # monitor.
- if self.is_send_to_remote():
- raise Exception(
- "assert_match() called on MultiResponseGdbRemoteEntry but state is set to send a query packet.")
-
- if self._end_matched:
- raise Exception(
- "assert_match() called on MultiResponseGdbRemoteEntry but end of query/response sequence has already been seen.")
-
- # Set up a context as needed.
- if not context:
- context = {}
-
- # Check if the packet matches the end condition.
- match = self._end_regex.match(actual_packet)
- if match:
- # We're done iterating.
- self._end_matched = True
- return context
-
- # Not done iterating - save the packet.
- context[self._save_key] = context.get(self._save_key, [])
- context[self._save_key].append(actual_packet)
-
- # Check for a runaway response cycle.
- if len(context[self._save_key]) >= self._runaway_response_count:
- raise Exception(
- "runaway query/response cycle detected: %d responses captured so far. Last response: %s" %
- (len(
- context[
- self._save_key]), context[
- self._save_key][
- -1]))
-
- # Flip the mode to send for generating the query.
- self._is_send_to_remote = True
- return context
-
-
-class MatchRemoteOutputEntry(GdbRemoteEntryBase):
- """Waits for output from the debug monitor to match a regex or time out.
-
- This entry type tries to match each time new gdb remote output is accumulated
- using a provided regex. If the output does not match the regex within the
- given timeframe, the command fails the playback session. If the regex does
- match, any capture fields are recorded in the context.
-
- Settings accepted from params:
-
- regex: required. Specifies a compiled regex object that must either succeed
- with re.match or re.search (see regex_mode below) within the given timeout
- (see timeout_seconds below) or cause the playback to fail.
-
- regex_mode: optional. Available values: "match" or "search". If "match", the entire
- stub output as collected so far must match the regex. If search, then the regex
- must match starting somewhere within the output text accumulated thus far.
- Default: "match" (i.e. the regex must match the entirety of the accumulated output
- buffer, so unexpected text will generally fail the match).
-
- capture: optional. If specified, is a dictionary of regex match group indices (should start
- with 1) to variable names that will store the capture group indicated by the
- index. For example, {1:"thread_id"} will store capture group 1's content in the
- context dictionary where "thread_id" is the key and the match group value is
- the value. The value stored off can be used later in a expect_captures expression.
- This arg only makes sense when regex is specified.
- """
-
- def __init__(self, regex=None, regex_mode="match", capture=None):
- self._regex = regex
- self._regex_mode = regex_mode
- self._capture = capture
- self._matched = False
-
- if not self._regex:
- raise Exception("regex cannot be None")
-
- if not self._regex_mode in ["match", "search"]:
- raise Exception(
- "unsupported regex mode \"{}\": must be \"match\" or \"search\"".format(
- self._regex_mode))
-
- def is_output_matcher(self):
- return True
-
- def is_send_to_remote(self):
- # This is always a "wait for remote" command.
- return False
-
- def is_consumed(self):
- return self._matched
-
- def assert_match(self, asserter, accumulated_output, context):
- # Validate args.
- if not accumulated_output:
- raise Exception("accumulated_output cannot be none")
- if not context:
- raise Exception("context cannot be none")
-
- # Validate that we haven't already matched.
- if self._matched:
- raise Exception(
- "invalid state - already matched, attempting to match again")
-
- # If we don't have any content yet, we don't match.
- if len(accumulated_output) < 1:
- return context
-
- # Check if we match
- if self._regex_mode == "match":
- match = self._regex.match(accumulated_output)
- elif self._regex_mode == "search":
- match = self._regex.search(accumulated_output)
- else:
- raise Exception(
- "Unexpected regex mode: {}".format(
- self._regex_mode))
-
- # If we don't match, wait to try again after next $O content, or time
- # out.
- if not match:
- # print("re pattern \"{}\" did not match against \"{}\"".format(self._regex.pattern, accumulated_output))
- return context
-
- # We do match.
- self._matched = True
- # print("re pattern \"{}\" matched against \"{}\"".format(self._regex.pattern, accumulated_output))
-
- # Collect up any captures into the context.
- if self._capture:
- # Handle captures.
- for group_index, var_name in list(self._capture.items()):
- capture_text = match.group(group_index)
- if not capture_text:
- raise Exception(
- "No content for group index {}".format(group_index))
- context[var_name] = capture_text
-
- return context
-
-
-class GdbRemoteTestSequence(object):
-
- _LOG_LINE_REGEX = re.compile(r'^.*(read|send)\s+packet:\s+(.+)$')
-
- def __init__(self, logger):
- self.entries = []
- self.logger = logger
-
- def add_log_lines(self, log_lines, remote_input_is_read):
- for line in log_lines:
- if isinstance(line, str):
- # Handle log line import
- # if self.logger:
- # self.logger.debug("processing log line: {}".format(line))
- match = self._LOG_LINE_REGEX.match(line)
- if match:
- playback_packet = match.group(2)
- direction = match.group(1)
- if _is_packet_lldb_gdbserver_input(
- direction, remote_input_is_read):
- # Handle as something to send to the remote debug monitor.
- # if self.logger:
- # self.logger.info("processed packet to send to remote: {}".format(playback_packet))
- self.entries.append(
- GdbRemoteEntry(
- is_send_to_remote=True,
- exact_payload=playback_packet))
- else:
- # Log line represents content to be expected from the remote debug monitor.
- # if self.logger:
- # self.logger.info("receiving packet from llgs, should match: {}".format(playback_packet))
- self.entries.append(
- GdbRemoteEntry(
- is_send_to_remote=False,
- exact_payload=playback_packet))
- else:
- raise Exception(
- "failed to interpret log line: {}".format(line))
- elif isinstance(line, dict):
- entry_type = line.get("type", "regex_capture")
- if entry_type == "regex_capture":
- # Handle more explicit control over details via dictionary.
- direction = line.get("direction", None)
- regex = line.get("regex", None)
- capture = line.get("capture", None)
- expect_captures = line.get("expect_captures", None)
-
- # Compile the regex.
- if regex and (isinstance(regex, str)):
- regex = re.compile(regex)
-
- if _is_packet_lldb_gdbserver_input(
- direction, remote_input_is_read):
- # Handle as something to send to the remote debug monitor.
- # if self.logger:
- # self.logger.info("processed dict sequence to send to remote")
- self.entries.append(
- GdbRemoteEntry(
- is_send_to_remote=True,
- regex=regex,
- capture=capture,
- expect_captures=expect_captures))
- else:
- # Log line represents content to be expected from the remote debug monitor.
- # if self.logger:
- # self.logger.info("processed dict sequence to match receiving from remote")
- self.entries.append(
- GdbRemoteEntry(
- is_send_to_remote=False,
- regex=regex,
- capture=capture,
- expect_captures=expect_captures))
- elif entry_type == "multi_response":
- self.entries.append(MultiResponseGdbRemoteEntry(line))
- elif entry_type == "output_match":
-
- regex = line.get("regex", None)
- # Compile the regex.
- if regex and (isinstance(regex, str)):
- regex = re.compile(regex, re.DOTALL)
-
- regex_mode = line.get("regex_mode", "match")
- capture = line.get("capture", None)
- self.entries.append(
- MatchRemoteOutputEntry(
- regex=regex,
- regex_mode=regex_mode,
- capture=capture))
- else:
- raise Exception("unknown entry type \"%s\"" % entry_type)
-
-
-def process_is_running(pid, unknown_value=True):
- """If possible, validate that the given pid represents a running process on the local system.
-
- Args:
-
- pid: an OS-specific representation of a process id. Should be an integral value.
-
- unknown_value: value used when we cannot determine how to check running local
- processes on the OS.
-
- Returns:
-
- If we can figure out how to check running process ids on the given OS:
- return True if the process is running, or False otherwise.
-
- If we don't know how to check running process ids on the given OS:
- return the value provided by the unknown_value arg.
- """
- if not isinstance(pid, six.integer_types):
- raise Exception(
- "pid must be an integral type (actual type: %s)" % str(
- type(pid)))
-
- process_ids = []
-
- if lldb.remote_platform:
- # Don't know how to get list of running process IDs on a remote
- # platform
- return unknown_value
- elif platform.system() in ['Darwin', 'Linux', 'FreeBSD', 'NetBSD']:
- # Build the list of running process ids
- output = subprocess.check_output(
- "ps ax | awk '{ print $1; }'", shell=True).decode("utf-8")
- text_process_ids = output.split('\n')[1:]
- # Convert text pids to ints
- process_ids = [int(text_pid)
- for text_pid in text_process_ids if text_pid != '']
- # elif {your_platform_here}:
- # fill in process_ids as a list of int type process IDs running on
- # the local system.
- else:
- # Don't know how to get list of running process IDs on this
- # OS, so return the "don't know" value.
- return unknown_value
-
- # Check if the pid is in the process_ids
- return pid in process_ids
-
-if __name__ == '__main__':
- EXE_PATH = get_lldb_server_exe()
- if EXE_PATH:
- print("lldb-server path detected: {}".format(EXE_PATH))
- else:
- print("lldb-server could not be found")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/main.cpp
deleted file mode 100644
index f1d46b85425..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/main.cpp
+++ /dev/null
@@ -1,370 +0,0 @@
-//===-- main.cpp ------------------------------------------------*- C++ -*-===//
-//
-// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
-// See https://llvm.org/LICENSE.txt for license information.
-// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
-//
-//===----------------------------------------------------------------------===//
-
-#include <atomic>
-#include <chrono>
-#include <cstdlib>
-#include <cstring>
-#include <errno.h>
-#include <inttypes.h>
-#include <memory>
-#include <mutex>
-#if !defined(_WIN32)
-#include <pthread.h>
-#include <signal.h>
-#include <unistd.h>
-#endif
-#include <setjmp.h>
-#include <stdint.h>
-#include <stdio.h>
-#include <string.h>
-#include <thread>
-#include <time.h>
-#include <vector>
-
-#if defined(__APPLE__)
-__OSX_AVAILABLE_STARTING(__MAC_10_6, __IPHONE_3_2)
-int pthread_threadid_np(pthread_t, __uint64_t *);
-#elif defined(__linux__)
-#include <sys/syscall.h>
-#elif defined(__NetBSD__)
-#include <lwp.h>
-#elif defined(_WIN32)
-#include <windows.h>
-#endif
-
-static const char *const RETVAL_PREFIX = "retval:";
-static const char *const SLEEP_PREFIX = "sleep:";
-static const char *const STDERR_PREFIX = "stderr:";
-static const char *const SET_MESSAGE_PREFIX = "set-message:";
-static const char *const PRINT_MESSAGE_COMMAND = "print-message:";
-static const char *const GET_DATA_ADDRESS_PREFIX = "get-data-address-hex:";
-static const char *const GET_STACK_ADDRESS_COMMAND = "get-stack-address-hex:";
-static const char *const GET_HEAP_ADDRESS_COMMAND = "get-heap-address-hex:";
-
-static const char *const GET_CODE_ADDRESS_PREFIX = "get-code-address-hex:";
-static const char *const CALL_FUNCTION_PREFIX = "call-function:";
-
-static const char *const THREAD_PREFIX = "thread:";
-static const char *const THREAD_COMMAND_NEW = "new";
-static const char *const THREAD_COMMAND_PRINT_IDS = "print-ids";
-static const char *const THREAD_COMMAND_SEGFAULT = "segfault";
-
-static const char *const PRINT_PID_COMMAND = "print-pid";
-
-static bool g_print_thread_ids = false;
-static std::mutex g_print_mutex;
-static bool g_threads_do_segfault = false;
-
-static std::mutex g_jump_buffer_mutex;
-static jmp_buf g_jump_buffer;
-static bool g_is_segfaulting = false;
-
-static char g_message[256];
-
-static volatile char g_c1 = '0';
-static volatile char g_c2 = '1';
-
-static void print_pid() {
-#if defined(_WIN32)
- fprintf(stderr, "PID: %d\n", ::GetCurrentProcessId());
-#else
- fprintf(stderr, "PID: %d\n", getpid());
-#endif
-}
-
-static void print_thread_id() {
-// Put in the right magic here for your platform to spit out the thread id (tid)
-// that debugserver/lldb-gdbserver would see as a TID. Otherwise, let the else
-// clause print out the unsupported text so that the unit test knows to skip
-// verifying thread ids.
-#if defined(__APPLE__)
- __uint64_t tid = 0;
- pthread_threadid_np(pthread_self(), &tid);
- printf("%" PRIx64, tid);
-#elif defined(__linux__)
- // This is a call to gettid() via syscall.
- printf("%" PRIx64, static_cast<uint64_t>(syscall(__NR_gettid)));
-#elif defined(__NetBSD__)
- // Technically lwpid_t is 32-bit signed integer
- printf("%" PRIx64, static_cast<uint64_t>(_lwp_self()));
-#elif defined(_WIN32)
- printf("%" PRIx64, static_cast<uint64_t>(::GetCurrentThreadId()));
-#else
- printf("{no-tid-support}");
-#endif
-}
-
-static void signal_handler(int signo) {
-#if defined(_WIN32)
- // No signal support on Windows.
-#else
- const char *signal_name = nullptr;
- switch (signo) {
- case SIGUSR1:
- signal_name = "SIGUSR1";
- break;
- case SIGSEGV:
- signal_name = "SIGSEGV";
- break;
- default:
- signal_name = nullptr;
- }
-
- // Print notice that we received the signal on a given thread.
- {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- if (signal_name)
- printf("received %s on thread id: ", signal_name);
- else
- printf("received signo %d (%s) on thread id: ", signo, strsignal(signo));
- print_thread_id();
- printf("\n");
- }
-
- // Reset the signal handler if we're one of the expected signal handlers.
- switch (signo) {
- case SIGSEGV:
- if (g_is_segfaulting) {
- // Fix up the pointer we're writing to. This needs to happen if nothing
- // intercepts the SIGSEGV (i.e. if somebody runs this from the command
- // line).
- longjmp(g_jump_buffer, 1);
- }
- break;
- case SIGUSR1:
- if (g_is_segfaulting) {
- // Fix up the pointer we're writing to. This is used to test gdb remote
- // signal delivery. A SIGSEGV will be raised when the thread is created,
- // switched out for a SIGUSR1, and then this code still needs to fix the
- // seg fault. (i.e. if somebody runs this from the command line).
- longjmp(g_jump_buffer, 1);
- }
- break;
- }
-
- // Reset the signal handler.
- sig_t sig_result = signal(signo, signal_handler);
- if (sig_result == SIG_ERR) {
- fprintf(stderr, "failed to set signal handler: errno=%d\n", errno);
- exit(1);
- }
-#endif
-}
-
-static void swap_chars() {
- g_c1 = '1';
- g_c2 = '0';
-
- g_c1 = '0';
- g_c2 = '1';
-}
-
-static void hello() {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("hello, world\n");
-}
-
-static void *thread_func(void *arg) {
- static std::atomic<int> s_thread_index(1);
- const int this_thread_index = s_thread_index++;
- if (g_print_thread_ids) {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("thread %d id: ", this_thread_index);
- print_thread_id();
- printf("\n");
- }
-
- if (g_threads_do_segfault) {
- // Sleep for a number of seconds based on the thread index.
- // TODO add ability to send commands to test exe so we can
- // handle timing more precisely. This is clunky. All we're
- // trying to do is add predictability as to the timing of
- // signal generation by created threads.
- int sleep_seconds = 2 * (this_thread_index - 1);
- std::this_thread::sleep_for(std::chrono::seconds(sleep_seconds));
-
- // Test creating a SEGV.
- {
- std::lock_guard<std::mutex> lock(g_jump_buffer_mutex);
- g_is_segfaulting = true;
- int *bad_p = nullptr;
- if (setjmp(g_jump_buffer) == 0) {
- // Force a seg fault signal on this thread.
- *bad_p = 0;
- } else {
- // Tell the system we're no longer seg faulting.
- // Used by the SIGUSR1 signal handler that we inject
- // in place of the SIGSEGV so it only tries to
- // recover from the SIGSEGV if this seg fault code
- // was in play.
- g_is_segfaulting = false;
- }
- }
-
- {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("thread ");
- print_thread_id();
- printf(": past SIGSEGV\n");
- }
- }
-
- int sleep_seconds_remaining = 60;
- std::this_thread::sleep_for(std::chrono::seconds(sleep_seconds_remaining));
-
- return nullptr;
-}
-
-int main(int argc, char **argv) {
- lldb_enable_attach();
-
- std::vector<std::thread> threads;
- std::unique_ptr<uint8_t[]> heap_array_up;
- int return_value = 0;
-
-#if !defined(_WIN32)
- // Set the signal handler.
- sig_t sig_result = signal(SIGALRM, signal_handler);
- if (sig_result == SIG_ERR) {
- fprintf(stderr, "failed to set SIGALRM signal handler: errno=%d\n", errno);
- exit(1);
- }
-
- sig_result = signal(SIGUSR1, signal_handler);
- if (sig_result == SIG_ERR) {
- fprintf(stderr, "failed to set SIGUSR1 handler: errno=%d\n", errno);
- exit(1);
- }
-
- sig_result = signal(SIGSEGV, signal_handler);
- if (sig_result == SIG_ERR) {
- fprintf(stderr, "failed to set SIGUSR1 handler: errno=%d\n", errno);
- exit(1);
- }
-#endif
-
- // Process command line args.
- for (int i = 1; i < argc; ++i) {
- if (std::strstr(argv[i], STDERR_PREFIX)) {
- // Treat remainder as text to go to stderr.
- fprintf(stderr, "%s\n", (argv[i] + strlen(STDERR_PREFIX)));
- } else if (std::strstr(argv[i], RETVAL_PREFIX)) {
- // Treat as the return value for the program.
- return_value = std::atoi(argv[i] + strlen(RETVAL_PREFIX));
- } else if (std::strstr(argv[i], SLEEP_PREFIX)) {
- // Treat as the amount of time to have this process sleep (in seconds).
- int sleep_seconds_remaining = std::atoi(argv[i] + strlen(SLEEP_PREFIX));
-
- // Loop around, sleeping until all sleep time is used up. Note that
- // signals will cause sleep to end early with the number of seconds
- // remaining.
- std::this_thread::sleep_for(
- std::chrono::seconds(sleep_seconds_remaining));
-
- } else if (std::strstr(argv[i], SET_MESSAGE_PREFIX)) {
- // Copy the contents after "set-message:" to the g_message buffer.
- // Used for reading inferior memory and verifying contents match
- // expectations.
- strncpy(g_message, argv[i] + strlen(SET_MESSAGE_PREFIX),
- sizeof(g_message));
-
- // Ensure we're null terminated.
- g_message[sizeof(g_message) - 1] = '\0';
-
- } else if (std::strstr(argv[i], PRINT_MESSAGE_COMMAND)) {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("message: %s\n", g_message);
- } else if (std::strstr(argv[i], GET_DATA_ADDRESS_PREFIX)) {
- volatile void *data_p = nullptr;
-
- if (std::strstr(argv[i] + strlen(GET_DATA_ADDRESS_PREFIX), "g_message"))
- data_p = &g_message[0];
- else if (std::strstr(argv[i] + strlen(GET_DATA_ADDRESS_PREFIX), "g_c1"))
- data_p = &g_c1;
- else if (std::strstr(argv[i] + strlen(GET_DATA_ADDRESS_PREFIX), "g_c2"))
- data_p = &g_c2;
-
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("data address: %p\n", data_p);
- } else if (std::strstr(argv[i], GET_HEAP_ADDRESS_COMMAND)) {
- // Create a byte array if not already present.
- if (!heap_array_up)
- heap_array_up.reset(new uint8_t[32]);
-
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("heap address: %p\n", heap_array_up.get());
-
- } else if (std::strstr(argv[i], GET_STACK_ADDRESS_COMMAND)) {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("stack address: %p\n", &return_value);
- } else if (std::strstr(argv[i], GET_CODE_ADDRESS_PREFIX)) {
- void (*func_p)() = nullptr;
-
- if (std::strstr(argv[i] + strlen(GET_CODE_ADDRESS_PREFIX), "hello"))
- func_p = hello;
- else if (std::strstr(argv[i] + strlen(GET_CODE_ADDRESS_PREFIX),
- "swap_chars"))
- func_p = swap_chars;
-
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("code address: %p\n", func_p);
- } else if (std::strstr(argv[i], CALL_FUNCTION_PREFIX)) {
- void (*func_p)() = nullptr;
-
- // Defaut to providing the address of main.
- if (std::strcmp(argv[i] + strlen(CALL_FUNCTION_PREFIX), "hello") == 0)
- func_p = hello;
- else if (std::strcmp(argv[i] + strlen(CALL_FUNCTION_PREFIX),
- "swap_chars") == 0)
- func_p = swap_chars;
- else {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("unknown function: %s\n",
- argv[i] + strlen(CALL_FUNCTION_PREFIX));
- }
- if (func_p)
- func_p();
- } else if (std::strstr(argv[i], THREAD_PREFIX)) {
- // Check if we're creating a new thread.
- if (std::strstr(argv[i] + strlen(THREAD_PREFIX), THREAD_COMMAND_NEW)) {
- threads.push_back(std::thread(thread_func, nullptr));
- } else if (std::strstr(argv[i] + strlen(THREAD_PREFIX),
- THREAD_COMMAND_PRINT_IDS)) {
- // Turn on thread id announcing.
- g_print_thread_ids = true;
-
- // And announce us.
- {
- std::lock_guard<std::mutex> lock(g_print_mutex);
- printf("thread 0 id: ");
- print_thread_id();
- printf("\n");
- }
- } else if (std::strstr(argv[i] + strlen(THREAD_PREFIX),
- THREAD_COMMAND_SEGFAULT)) {
- g_threads_do_segfault = true;
- } else {
- // At this point we don't do anything else with threads.
- // Later use thread index and send command to thread.
- }
- } else if (std::strstr(argv[i], PRINT_PID_COMMAND)) {
- print_pid();
- } else {
- // Treat the argument as text for stdout.
- printf("%s\n", argv[i]);
- }
- }
-
- // If we launched any threads, join them
- for (std::vector<std::thread>::iterator it = threads.begin();
- it != threads.end(); ++it)
- it->join();
-
- return return_value;
-}
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/Makefile
deleted file mode 100644
index 314f1cb2f07..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/Makefile
+++ /dev/null
@@ -1,5 +0,0 @@
-LEVEL = ../../../make
-
-CXX_SOURCES := main.cpp
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/TestPlatformProcessConnect.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/TestPlatformProcessConnect.py
deleted file mode 100644
index aa4b3dee792..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/TestPlatformProcessConnect.py
+++ /dev/null
@@ -1,96 +0,0 @@
-from __future__ import print_function
-
-import time
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestPlatformProcessConnect(gdbremote_testcase.GdbRemoteTestCaseBase):
- mydir = TestBase.compute_mydir(__file__)
-
- @llgs_test
- @no_debug_info_test
- @skipIf(remote=False)
- @expectedFailureAll(hostoslist=["windows"], triple='.*-android')
- def test_platform_process_connect(self):
- self.build()
- self.init_llgs_test(False)
-
- working_dir = lldb.remote_platform.GetWorkingDirectory()
- src = lldb.SBFileSpec(self.getBuildArtifact("a.out"))
- dest = lldb.SBFileSpec(os.path.join(working_dir, "a.out"))
- err = lldb.remote_platform.Put(src, dest)
- if err.Fail():
- raise RuntimeError(
- "Unable copy '%s' to '%s'.\n>>> %s" %
- (f, wd, err.GetCString()))
-
- m = re.search("^(.*)://([^:/]*)", configuration.lldb_platform_url)
- protocol = m.group(1)
- hostname = m.group(2)
- unix_protocol = protocol.startswith("unix-")
- if unix_protocol:
- p = re.search("^(.*)-connect", protocol)
- path = lldbutil.join_remote_paths(configuration.lldb_platform_working_dir,
- self.getBuildDirBasename(), "platform-%d.sock" % int(time.time()))
- listen_url = "%s://%s" % (p.group(1), path)
- else:
- listen_url = "*:0"
-
- port_file = "%s/port" % working_dir
- commandline_args = [
- "platform",
- "--listen",
- listen_url,
- "--socket-file",
- port_file,
- "--",
- "%s/a.out" %
- working_dir,
- "foo"]
- self.spawnSubprocess(
- self.debug_monitor_exe,
- commandline_args,
- install_remote=False)
- self.addTearDownHook(self.cleanupSubprocesses)
-
- socket_id = lldbutil.wait_for_file_on_target(self, port_file)
-
- new_debugger = lldb.SBDebugger.Create()
- new_debugger.SetAsync(False)
-
- def del_debugger(new_debugger=new_debugger):
- del new_debugger
- self.addTearDownHook(del_debugger)
-
- new_platform = lldb.SBPlatform(lldb.remote_platform.GetName())
- new_debugger.SetSelectedPlatform(new_platform)
- new_interpreter = new_debugger.GetCommandInterpreter()
-
- if unix_protocol:
- connect_url = "%s://%s%s" % (protocol, hostname, socket_id)
- else:
- connect_url = "%s://%s:%s" % (protocol, hostname, socket_id)
-
- command = "platform connect %s" % (connect_url)
- result = lldb.SBCommandReturnObject()
- new_interpreter.HandleCommand(command, result)
- self.assertTrue(
- result.Succeeded(),
- "platform process connect failed: %s" %
- result.GetOutput())
-
- target = new_debugger.GetSelectedTarget()
- process = target.GetProcess()
- thread = process.GetThreadAtIndex(0)
-
- breakpoint = target.BreakpointCreateByName("main")
- process.Continue()
-
- frame = thread.GetFrameAtIndex(0)
- self.assertEqual(frame.GetFunction().GetName(), "main")
- self.assertEqual(frame.FindVariable("argc").GetValueAsSigned(), 2)
- process.Continue()
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/main.cpp
deleted file mode 100644
index c7ebe0759a4..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/platform-process-connect/main.cpp
+++ /dev/null
@@ -1,6 +0,0 @@
-#include <cstdio>
-
-int main(int argc, char **argv) {
- printf("argc: %d\n", argc);
- return argv[0][0];
-}
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/Makefile
deleted file mode 100644
index 314f1cb2f07..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/Makefile
+++ /dev/null
@@ -1,5 +0,0 @@
-LEVEL = ../../../make
-
-CXX_SOURCES := main.cpp
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/TestGdbRemoteGPacket.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/TestGdbRemoteGPacket.py
deleted file mode 100644
index e13daeb6d9a..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/TestGdbRemoteGPacket.py
+++ /dev/null
@@ -1,153 +0,0 @@
-from __future__ import print_function
-
-
-import gdbremote_testcase
-import textwrap
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-def _extract_register_value(reg_info, reg_bank, byte_order, bytes_per_entry=8):
- reg_offset = int(reg_info["offset"])*2
- reg_byte_size = int(2 * int(reg_info["bitsize"]) / 8)
- # Create slice with the contents of the register.
- reg_slice = reg_bank[reg_offset:reg_offset+reg_byte_size]
-
- reg_value = []
- # Wrap slice according to bytes_per_entry.
- for entry in textwrap.wrap(reg_slice, 2 * bytes_per_entry):
- # Invert the bytes order if target uses little-endian.
- if byte_order == lldb.eByteOrderLittle:
- entry = "".join(reversed([entry[i:i+2] for i in range(0,
- len(entry),2)]))
- reg_value.append("0x" + entry)
-
- return reg_value
-
-
-class TestGdbRemoteGPacket(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def run_test_g_packet(self):
- self.build()
- self.prep_debug_monitor_and_inferior()
- self.test_sequence.add_log_lines(
- ["read packet: $g#67",
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "register_bank"}}],
- True)
- self.connect_to_debug_monitor()
- context = self.expect_gdbremote_sequence()
- register_bank = context.get("register_bank")
- self.assertTrue(register_bank[0] != 'E')
-
- self.test_sequence.add_log_lines(
- ["read packet: $G" + register_bank + "#00",
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "G_reply"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertTrue(context.get("G_reply")[0] != 'E')
-
- @skipIfOutOfTreeDebugserver
- @debugserver_test
- @skipIfDarwinEmbedded
- def test_g_packet_debugserver(self):
- self.init_debugserver_test()
- self.run_test_g_packet()
-
- @skipIf(archs=no_match(["x86_64"]))
- def g_returns_correct_data(self, with_suffix):
- procs = self.prep_debug_monitor_and_inferior()
-
- self.add_register_info_collection_packets()
- if with_suffix:
- self.add_thread_suffix_request_packets()
- self.add_threadinfo_collection_packets()
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Gather register info.
- reg_infos = self.parse_register_info_packets(context)
- self.assertIsNotNone(reg_infos)
- self.add_lldb_register_index(reg_infos)
- # Index register info entries by name.
- reg_infos = {info['name']: info for info in reg_infos}
-
- # Gather thread info.
- if with_suffix:
- threads = self.parse_threadinfo_packets(context)
- self.assertIsNotNone(threads)
- thread_id = threads[0]
- self.assertIsNotNone(thread_id)
- else:
- thread_id = None
-
- # Send vCont packet to resume the inferior.
- self.test_sequence.add_log_lines(["read packet: $vCont;c#a8",
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2}).*#[0-9a-fA-F]{2}$",
- "capture": {1: "hex_exit_code"}},
- ],
- True)
-
- # Send g packet to retrieve the register bank
- if thread_id:
- g_request = "read packet: $g;thread:{:x}#00".format(thread_id)
- else:
- g_request = "read packet: $g#00"
- self.test_sequence.add_log_lines(
- [g_request,
- {"direction": "send", "regex": r"^\$(.+)#[0-9a-fA-F]{2}$",
- "capture": {1: "register_bank"}}],
- True)
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
- reg_bank = context.get("register_bank")
- self.assertTrue(reg_bank[0] != 'E')
-
- byte_order = self.get_target_byte_order()
- get_reg_value = lambda reg_name : _extract_register_value(
- reg_infos[reg_name], reg_bank, byte_order)
-
- self.assertEqual(['0x0102030405060708'], get_reg_value('r8'))
- self.assertEqual(['0x1112131415161718'], get_reg_value('r9'))
- self.assertEqual(['0x2122232425262728'], get_reg_value('r10'))
- self.assertEqual(['0x3132333435363738'], get_reg_value('r11'))
- self.assertEqual(['0x4142434445464748'], get_reg_value('r12'))
- self.assertEqual(['0x5152535455565758'], get_reg_value('r13'))
- self.assertEqual(['0x6162636465666768'], get_reg_value('r14'))
- self.assertEqual(['0x7172737475767778'], get_reg_value('r15'))
-
- self.assertEqual(
- ['0x020406080a0c0e01', '0x030507090b0d0f00'], get_reg_value('xmm8'))
- self.assertEqual(
- ['0x121416181a1c1e11', '0x131517191b1d1f10'], get_reg_value('xmm9'))
- self.assertEqual(
- ['0x222426282a2c2e21', '0x232527292b2d2f20'], get_reg_value('xmm10'))
- self.assertEqual(
- ['0x323436383a3c3e31', '0x333537393b3d3f30'], get_reg_value('xmm11'))
- self.assertEqual(
- ['0x424446484a4c4e41', '0x434547494b4d4f40'], get_reg_value('xmm12'))
- self.assertEqual(
- ['0x525456585a5c5e51', '0x535557595b5d5f50'], get_reg_value('xmm13'))
- self.assertEqual(
- ['0x626466686a6c6e61', '0x636567696b6d6f60'], get_reg_value('xmm14'))
- self.assertEqual(
- ['0x727476787a7c7e71', '0x737577797b7d7f70'], get_reg_value('xmm15'))
-
- @llgs_test
- def test_g_returns_correct_data_with_suffix_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.g_returns_correct_data(True)
-
- @llgs_test
- def test_g_returns_correct_data_no_suffix_llgs(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- self.g_returns_correct_data(False)
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/main.cpp
deleted file mode 100644
index 32eda6d3c55..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/register-reading/main.cpp
+++ /dev/null
@@ -1,54 +0,0 @@
-#include <cstdint>
-
-struct alignas(16) xmm_t {
- uint64_t a, b;
-};
-
-int main() {
- uint64_t r8 = 0x0102030405060708;
- uint64_t r9 = 0x1112131415161718;
- uint64_t r10 = 0x2122232425262728;
- uint64_t r11 = 0x3132333435363738;
- uint64_t r12 = 0x4142434445464748;
- uint64_t r13 = 0x5152535455565758;
- uint64_t r14 = 0x6162636465666768;
- uint64_t r15 = 0x7172737475767778;
-
- xmm_t xmm8 = {0x020406080A0C0E01, 0x030507090B0D0F00};
- xmm_t xmm9 = {0x121416181A1C1E11, 0x131517191B1D1F10};
- xmm_t xmm10 = {0x222426282A2C2E21, 0x232527292B2D2F20};
- xmm_t xmm11 = {0x323436383A3C3E31, 0x333537393B3D3F30};
- xmm_t xmm12 = {0x424446484A4C4E41, 0x434547494B4D4F40};
- xmm_t xmm13 = {0x525456585A5C5E51, 0x535557595B5D5F50};
- xmm_t xmm14 = {0x626466686A6C6E61, 0x636567696B6D6F60};
- xmm_t xmm15 = {0x727476787A7C7E71, 0x737577797B7D7F70};
-
- asm volatile("movq %0, %%r8\n\t"
- "movq %1, %%r9\n\t"
- "movq %2, %%r10\n\t"
- "movq %3, %%r11\n\t"
- "movq %4, %%r12\n\t"
- "movq %5, %%r13\n\t"
- "movq %6, %%r14\n\t"
- "movq %7, %%r15\n\t"
- "\n\t"
- "movaps %8, %%xmm8\n\t"
- "movaps %9, %%xmm9\n\t"
- "movaps %10, %%xmm10\n\t"
- "movaps %11, %%xmm11\n\t"
- "movaps %12, %%xmm12\n\t"
- "movaps %13, %%xmm13\n\t"
- "movaps %14, %%xmm14\n\t"
- "movaps %15, %%xmm15\n\t"
- "\n\t"
- "int3"
- :
- : "g"(r8), "g"(r9), "g"(r10), "g"(r11), "g"(r12), "g"(r13),
- "g"(r14), "g"(r15), "m"(xmm8), "m"(xmm9), "m"(xmm10),
- "m"(xmm11), "m"(xmm12), "m"(xmm13), "m"(xmm14), "m"(xmm15)
- : "%r8", "%r9", "%r10", "%r11", "%r12", "%r13", "%r14", "%r15",
- "%xmm8", "%xmm9", "%xmm10", "%xmm11", "%xmm12", "%xmm13",
- "%xmm14", "%xmm15");
-
- return 0;
-}
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/Makefile
deleted file mode 100644
index 314f1cb2f07..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/Makefile
+++ /dev/null
@@ -1,5 +0,0 @@
-LEVEL = ../../../make
-
-CXX_SOURCES := main.cpp
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/TestGdbRemote_QPassSignals.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/TestGdbRemote_QPassSignals.py
deleted file mode 100644
index 7105bcb078b..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/TestGdbRemote_QPassSignals.py
+++ /dev/null
@@ -1,115 +0,0 @@
-# This test makes sure that lldb-server supports and properly handles
-# QPassSignals GDB protocol package.
-from __future__ import print_function
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-class TestGdbRemote_QPassSignals(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def expect_signal(self, expected_signo):
- self.test_sequence.add_log_lines(["read packet: $vCont;c#a8",
- {"direction": "send",
- "regex": r"^\$T([0-9a-fA-F]{2}).*#[0-9a-fA-F]{2}$",
- "capture": {1: "hex_exit_code"}},
- ],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- hex_exit_code = context.get("hex_exit_code")
- self.assertIsNotNone(hex_exit_code)
- self.assertEqual(int(hex_exit_code, 16), expected_signo)
-
- def expect_exit_code(self, exit_code):
- self.test_sequence.add_log_lines(
- ["read packet: $vCont;c#a8",
- "send packet: $W{0:02x}#00".format(exit_code)],
- True)
- self.expect_gdbremote_sequence()
-
-
- def ignore_signals(self, signals):
- def signal_name_to_hex(signame):
- return format(lldbutil.get_signal_number(signame), 'x')
- signals_str = ";".join(map(signal_name_to_hex, signals))
-
- self.test_sequence.add_log_lines(["read packet: $QPassSignals:"
- + signals_str + " #00",
- "send packet: $OK#00"],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- @llgs_test
- @skipUnlessPlatform(["linux", "android"])
- def test_q_pass_signals(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- procs = self.prep_debug_monitor_and_inferior()
- expected_signals = ["SIGSEGV",
- "SIGALRM", "SIGFPE", "SIGBUS", "SIGINT", "SIGHUP"]
- signals_to_ignore = ["SIGUSR1", "SIGUSR2"]
- self.ignore_signals(signals_to_ignore)
- for signal_name in expected_signals:
- signo = lldbutil.get_signal_number(signal_name)
- self.expect_signal(signo)
- self.expect_exit_code(len(signals_to_ignore))
-
- @llgs_test
- @skipUnlessPlatform(["linux", "android"])
- def test_change_signals_at_runtime(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- procs = self.prep_debug_monitor_and_inferior()
- expected_signals = ["SIGSEGV", "SIGUSR1", "SIGUSR2",
- "SIGALRM", "SIGHUP"]
- signals_to_ignore = ["SIGFPE", "SIGBUS", "SIGINT"]
-
- for signal_name in expected_signals:
- signo = lldbutil.get_signal_number(signal_name)
- self.expect_signal(signo)
- if signal_name == "SIGALRM":
- self.ignore_signals(signals_to_ignore)
- self.expect_exit_code(len(signals_to_ignore))
-
- @llgs_test
- def test_default_signals_behavior(self):
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- procs = self.prep_debug_monitor_and_inferior()
- expected_signals = ["SIGSEGV", "SIGUSR1", "SIGUSR2",
- "SIGALRM", "SIGFPE", "SIGBUS", "SIGINT", "SIGHUP"]
- for signal_name in expected_signals:
- signo = lldbutil.get_signal_number(signal_name)
- self.expect_signal(signo)
- self.expect_exit_code(0)
-
-
- @llgs_test
- @skipUnlessPlatform(["linux", "android"])
- def test_support_q_pass_signals(self):
- self.init_llgs_test()
- self.build()
-
- # Start up the stub and start/prep the inferior.
- self.set_inferior_startup_launch()
- procs = self.prep_debug_monitor_and_inferior()
- self.add_qSupported_packets()
-
- # Run the packet stream.
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- # Retrieve the qSupported features and check QPassSignals+
- supported_dict = self.parse_qSupported_response(context)
- self.assertEqual(supported_dict["QPassSignals"], "+")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/main.cpp
deleted file mode 100644
index fe33c291958..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/signal-filtering/main.cpp
+++ /dev/null
@@ -1,36 +0,0 @@
-//===-- main.cpp ------------------------------------------------*- C++ -*-===//
-//
-// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
-// See https://llvm.org/LICENSE.txt for license information.
-// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
-//
-//===----------------------------------------------------------------------===//
-
-#include <signal.h>
-#include <stdio.h>
-#include <vector>
-
-static int signal_counter = 0;
-
-static void count_signal(int signo) {
- ++signal_counter;
- printf("Signal %d\n", signo);
-}
-
-static void raise_signals() {
- std::vector<int> signals(
- {SIGSEGV, SIGUSR1, SIGUSR2, SIGALRM, SIGFPE, SIGBUS, SIGINT, SIGHUP});
-
- for (int signal_num : signals) {
- signal(signal_num, count_signal);
- }
-
- for (int signal_num : signals) {
- raise(signal_num);
- }
-}
-
-int main() {
- raise_signals();
- return signal_counter;
-}
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/socket_packet_pump.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/socket_packet_pump.py
deleted file mode 100644
index 958d6449b51..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/socket_packet_pump.py
+++ /dev/null
@@ -1,198 +0,0 @@
-
-from __future__ import print_function
-
-
-import re
-import select
-import threading
-import traceback
-import codecs
-
-from six.moves import queue
-from lldbsuite.support import seven
-
-
-def _handle_output_packet_string(packet_contents):
- if (not packet_contents) or (len(packet_contents) < 1):
- return None
- elif packet_contents[0] != "O":
- return None
- elif packet_contents == "OK":
- return None
- else:
- return seven.unhexlify(packet_contents[1:])
-
-
-def _dump_queue(the_queue):
- while not the_queue.empty():
- print(codecs.encode(the_queue.get(True), "string_escape"))
- print("\n")
-
-
-class PumpQueues(object):
-
- def __init__(self):
- self._output_queue = queue.Queue()
- self._packet_queue = queue.Queue()
-
- def output_queue(self):
- return self._output_queue
-
- def packet_queue(self):
- return self._packet_queue
-
- def verify_queues_empty(self):
- # Warn if there is any content left in any of the queues.
- # That would represent unmatched packets.
- if not self.output_queue().empty():
- print("warning: output queue entries still exist:")
- _dump_queue(self.output_queue())
- print("from here:")
- traceback.print_stack()
-
- if not self.packet_queue().empty():
- print("warning: packet queue entries still exist:")
- _dump_queue(self.packet_queue())
- print("from here:")
- traceback.print_stack()
-
-
-class SocketPacketPump(object):
- """A threaded packet reader that partitions packets into two streams.
-
- All incoming $O packet content is accumulated with the current accumulation
- state put into the OutputQueue.
-
- All other incoming packets are placed in the packet queue.
-
- A select thread can be started and stopped, and runs to place packet
- content into the two queues.
- """
-
- _GDB_REMOTE_PACKET_REGEX = re.compile(r'^\$([^\#]*)#[0-9a-fA-F]{2}')
-
- def __init__(self, pump_socket, pump_queues, logger=None):
- if not pump_socket:
- raise Exception("pump_socket cannot be None")
-
- self._thread = None
- self._stop_thread = False
- self._socket = pump_socket
- self._logger = logger
- self._receive_buffer = ""
- self._accumulated_output = ""
- self._pump_queues = pump_queues
-
- def __enter__(self):
- """Support the python 'with' statement.
-
- Start the pump thread."""
- self.start_pump_thread()
- return self
-
- def __exit__(self, exit_type, value, the_traceback):
- """Support the python 'with' statement.
-
- Shut down the pump thread."""
- self.stop_pump_thread()
-
- def start_pump_thread(self):
- if self._thread:
- raise Exception("pump thread is already running")
- self._stop_thread = False
- self._thread = threading.Thread(target=self._run_method)
- self._thread.start()
-
- def stop_pump_thread(self):
- self._stop_thread = True
- if self._thread:
- self._thread.join()
-
- def _process_new_bytes(self, new_bytes):
- if not new_bytes:
- return
- if len(new_bytes) < 1:
- return
-
- # Add new bytes to our accumulated unprocessed packet bytes.
- self._receive_buffer += new_bytes
-
- # Parse fully-formed packets into individual packets.
- has_more = len(self._receive_buffer) > 0
- while has_more:
- if len(self._receive_buffer) <= 0:
- has_more = False
- # handle '+' ack
- elif self._receive_buffer[0] == "+":
- self._pump_queues.packet_queue().put("+")
- self._receive_buffer = self._receive_buffer[1:]
- if self._logger:
- self._logger.debug(
- "parsed packet from stub: +\n" +
- "new receive_buffer: {}".format(
- self._receive_buffer))
- else:
- packet_match = self._GDB_REMOTE_PACKET_REGEX.match(
- self._receive_buffer)
- if packet_match:
- # Our receive buffer matches a packet at the
- # start of the receive buffer.
- new_output_content = _handle_output_packet_string(
- packet_match.group(1))
- if new_output_content:
- # This was an $O packet with new content.
- self._accumulated_output += new_output_content
- self._pump_queues.output_queue().put(self._accumulated_output)
- else:
- # Any packet other than $O.
- self._pump_queues.packet_queue().put(packet_match.group(0))
-
- # Remove the parsed packet from the receive
- # buffer.
- self._receive_buffer = self._receive_buffer[
- len(packet_match.group(0)):]
- if self._logger:
- self._logger.debug(
- "parsed packet from stub: " +
- packet_match.group(0))
- self._logger.debug(
- "new receive_buffer: " +
- self._receive_buffer)
- else:
- # We don't have enough in the receive bufferto make a full
- # packet. Stop trying until we read more.
- has_more = False
-
- def _run_method(self):
- self._receive_buffer = ""
- self._accumulated_output = ""
-
- if self._logger:
- self._logger.info("socket pump starting")
-
- # Keep looping around until we're asked to stop the thread.
- while not self._stop_thread:
- can_read, _, _ = select.select([self._socket], [], [], 0)
- if can_read and self._socket in can_read:
- try:
- new_bytes = seven.bitcast_to_string(self._socket.recv(4096))
- if self._logger and new_bytes and len(new_bytes) > 0:
- self._logger.debug(
- "pump received bytes: {}".format(new_bytes))
- except:
- # Likely a closed socket. Done with the pump thread.
- if self._logger:
- self._logger.debug(
- "socket read failed, stopping pump read thread\n" +
- traceback.format_exc(3))
- break
- self._process_new_bytes(new_bytes)
-
- if self._logger:
- self._logger.info("socket pump exiting")
-
- def get_accumulated_output(self):
- return self._accumulated_output
-
- def get_receive_buffer(self):
- return self._receive_buffer
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/test/test_lldbgdbserverutils.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/test/test_lldbgdbserverutils.py
deleted file mode 100644
index dc52f244aa4..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/test/test_lldbgdbserverutils.py
+++ /dev/null
@@ -1,65 +0,0 @@
-from __future__ import print_function
-
-
-import unittest2
-import os.path
-import re
-import sys
-
-from lldbgdbserverutils import *
-
-
-class TestLldbGdbServerUtils(unittest2.TestCase):
-
- def test_entry_exact_payload_match(self):
- entry = GdbRemoteEntry(is_send_to_remote=False, exact_payload="$OK#9a")
- entry.assert_match(self, "$OK#9a")
-
- def test_entry_exact_payload_match_ignores_checksum(self):
- entry = GdbRemoteEntry(is_send_to_remote=False, exact_payload="$OK#9a")
- entry.assert_match(self, "$OK#00")
-
- def test_entry_creates_context(self):
- entry = GdbRemoteEntry(is_send_to_remote=False, exact_payload="$OK#9a")
- context = entry.assert_match(self, "$OK#9a")
- self.assertIsNotNone(context)
-
- def test_entry_regex_matches(self):
- entry = GdbRemoteEntry(
- is_send_to_remote=False,
- regex=re.compile(r"^\$QC([0-9a-fA-F]+)#"),
- capture={
- 1: "thread_id"})
- context = entry.assert_match(self, "$QC980#00")
-
- def test_entry_regex_saves_match(self):
- entry = GdbRemoteEntry(
- is_send_to_remote=False,
- regex=re.compile(r"^\$QC([0-9a-fA-F]+)#"),
- capture={
- 1: "thread_id"})
- context = entry.assert_match(self, "$QC980#00")
- self.assertEqual(context["thread_id"], "980")
-
- def test_entry_regex_expect_captures_success(self):
- context = {"thread_id": "980"}
- entry = GdbRemoteEntry(
- is_send_to_remote=False,
- regex=re.compile(r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+)"),
- expect_captures={
- 2: "thread_id"})
- entry.assert_match(self, "$T11thread:980;", context=context)
-
- def test_entry_regex_expect_captures_raises_on_fail(self):
- context = {"thread_id": "980"}
- entry = GdbRemoteEntry(
- is_send_to_remote=False,
- regex=re.compile(r"^\$T([0-9a-fA-F]{2})thread:([0-9a-fA-F]+)"),
- expect_captures={
- 2: "thread_id"})
- try:
- entry.assert_match(self, "$T11thread:970;", context=context)
- self.fail()
- except AssertionError:
- # okay
- return None
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/Makefile b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/Makefile
deleted file mode 100644
index 8817fff55e8..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/Makefile
+++ /dev/null
@@ -1,6 +0,0 @@
-LEVEL = ../../../make
-
-ENABLE_THREADS := YES
-CXX_SOURCES := main.cpp
-
-include $(LEVEL)/Makefile.rules
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/TestGdbRemoteThreadName.py b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/TestGdbRemoteThreadName.py
deleted file mode 100644
index 5bfcd660c2a..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/TestGdbRemoteThreadName.py
+++ /dev/null
@@ -1,41 +0,0 @@
-from __future__ import print_function
-
-import gdbremote_testcase
-from lldbsuite.test.decorators import *
-from lldbsuite.test.lldbtest import *
-from lldbsuite.test import lldbutil
-
-
-class TestGdbRemoteThreadName(gdbremote_testcase.GdbRemoteTestCaseBase):
-
- mydir = TestBase.compute_mydir(__file__)
-
- def run_and_check_name(self, expected_name):
- self.test_sequence.add_log_lines(["read packet: $vCont;c#a8",
- {"direction": "send",
- "regex":
- r"^\$T([0-9a-fA-F]{2})([^#]+)#[0-9a-fA-F]{2}$",
- "capture": {
- 1: "signal",
- 2: "key_vals_text"}},
- ],
- True)
-
- context = self.expect_gdbremote_sequence()
- self.assertIsNotNone(context)
-
- sigint = lldbutil.get_signal_number("SIGINT")
- self.assertEqual(sigint, int(context.get("signal"), 16))
- kv_dict = self.parse_key_val_dict(context.get("key_vals_text"))
- self.assertEqual(expected_name, kv_dict.get("name"))
-
- @llgs_test
- def test(self):
- """ Make sure lldb-server can retrieve inferior thread name"""
- self.init_llgs_test()
- self.build()
- self.set_inferior_startup_launch()
- procs = self.prep_debug_monitor_and_inferior()
-
- self.run_and_check_name("hello world")
- self.run_and_check_name("goodbye world")
diff --git a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/main.cpp b/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/main.cpp
deleted file mode 100644
index 0403031143b..00000000000
--- a/lldb/packages/Python/lldbsuite/test/tools/lldb-server/thread-name/main.cpp
+++ /dev/null
@@ -1,22 +0,0 @@
-#include <pthread.h>
-#include <signal.h>
-
-void set_thread_name(const char *name) {
-#if defined(__APPLE__)
- ::pthread_setname_np(name);
-#elif defined(__FreeBSD__)
- ::pthread_set_name_np(::pthread_self(), name);
-#elif defined(__linux__)
- ::pthread_setname_np(::pthread_self(), name);
-#elif defined(__NetBSD__)
- ::pthread_setname_np(::pthread_self(), "%s", name);
-#endif
-}
-
-int main() {
- set_thread_name("hello world");
- raise(SIGINT);
- set_thread_name("goodbye world");
- raise(SIGINT);
- return 0;
-}
OpenPOWER on IntegriCloud