depending on the distro you could use a .deb or .tar.gz instead of binaries and then install it with your package manager
Linux
From Wikipedia, the free encyclopedia
Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).
Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.
Rules
- Posts must be relevant to operating systems running the Linux kernel. GNU/Linux or otherwise.
- No misinformation
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
Community icon by AlpΓ‘r-Etele MΓ©der, licensed under CC BY 3.0
Why not use a live ISO version of something and boot it from a USB, if you need a full set of network troubleshooting tools, the Kali Linux Everything ISO for example will definitely have everything.
this sounds like a good idea - I'll try that :))
Yeah, that should work. ldd "$(command -v "$cmd")"
will list the dynamic dependencies for $cmd
, so you can find those (probably) in /lib
and /usr/lib
; I'm not familiar enough with the dynamic library loading process to give you the specifics. I would put the binaries in /usr/local/bin
and the libraries in /usr/local/lib
; but you could also modify path variables to point to the usb drive. Ideally you could find statically linked versions somewhere, so you don't have to mess with the libraries.
Alternatively, most package managers have commands to download packages; then you can copy the package cache over to the new machine and install them that way. If the commands are common enough, you could download one of the bigger install media and add its package repo to your machine. These of course are distribution specific processes.
Finally, you could get a cheap USB ethernet adapter and connect to the internet that way. On newegg most of these products will have at least one review saying whether they work on linux.
If you have an android phone, you can plug it in via USB and enable USB Internet tethering, which will give you working internet access on your machine to do the Wi-Fi debugging with.
In general, no. Better way is to download packages with that tools from your distro repository, transfer them via flash key and install. You also have to download dependencies, but CLI tools usually have few of them and there are good chances they are already installed.
Depends on the tools. If they're statically compiled, it should be fine. If they aren't, it might still be fine if the distro and versions are similar. But what you want is statically compiled binaries.
It'll need to be the same architecture (ARM -> ARM good, AMD -> ARM bad), and check each tool on your working computer with ldd
; the fewer lib dependencies, the better.
Scripting languages are probably not worth messing with. Even if you have a running interpreter on the broken machine, scripting languages tend to lean heavily on third party libs, which may not be installed. The exception are ba/sh scripts, which have a good chance of using only commonly installed commands (why else use bash?).
When I had no (useful) Internet where I was living a few years ago I would save a list of packages to download from Synaptic to a drive and then when I was somewhere I could I would download them, then when I got home I could plug in the drive and update/install them.
A statically linked copy of busybox probably has 90% of the functionally you need
If the package manager on your old PC is keeping copies of everything it installs, just copy all of those packages over and go through the package manager on the new PC. Look under /var/cache