Hey folks!
Apologies if this has been answered before and my forum search skill are lacking. I'm trying to confirm my understanding of the behavior of "-speed auto" parameter of "JLinkGDBServerCL". Is it basically equal to "best possible speed given the debug adapter and target MCU limitations"? What's the heuristic that JLinkGDBServerCL would use? Would it set the interface / JTAG / SWO speed to maximum supported by the adapter and then gradually slower it if it encounters communication errors?