Want to improve this cheat sheet? See the Contributing section!
"With Docker, developers can build any app in any language using any toolchain. “Dockerized” apps are completely portable and can run anywhere - colleagues’ OS X and Windows laptops, QA servers running Ubuntu in the cloud, and production data center VMs running Red Hat.
Developers can get going quickly by starting with one of the 13,000+ apps available on Docker Hub. Docker manages and tracks changes and dependencies, making it easier for sysadmins to understand how the apps that developers build work. And with Docker Hub, developers can automate their build pipeline and share artifacts with collaborators through public or private repositories.
Docker helps developers build and ship higher-quality applications, faster." -- What is Docker
I use Oh My Zsh with the Docker plugin for autocompletion of docker commands. YMMV.
The 3.10.x kernel is the minimum requirement for Docker.
10.8 “Mountain Lion” or newer is required.
Hyper-V must be enabled in BIOS
VT-D must also be enabled if available (Intel Processors).
Windows Server 2016 is the minimum version required to install docker and docker-compose. Limitations exist on this version, such as multiple virtual networks and Linux containers. Windows Server 2019 and later are recommended.
Run this quick and easy install script provided by Docker:
curl -sSL https://get.docker.com/ | sh
If you're not willing to run a random shell script, please see the installation instructions for your distribution.
If you are a complete Docker newbie, you should follow the series of tutorials now.
Download and install Docker Community Edition. if you have Homebrew-Cask, just type brew cask install docker
. Or Download and install Docker Toolbox. Docker For Mac is nice, but it's not quite as finished as the VirtualBox install. See the comparison.
NOTE Docker Toolbox is legacy. You should to use Docker Community Edition, See Docker Toolbox.
Once you've installed Docker Community Edition, click the docker icon in Launchpad. Then start up a container:
docker run hello-world
That's it, you have a running Docker container.
If you are a complete Docker newbie, you should probably follow the series of tutorials now.
Instructions to install Docker Desktop for Windows can be found here
Once installed, open powershell as administrator and run:
# Display the version of docker installed:
docker version
# Pull, create, and run 'hello-world':
docker run hello-world
To continue with this cheat sheet, right click the Docker icon in the system tray, and go to settings. In order to mount volumes, the C:/ drive will need to be enabled in the settings to that information can be passed into the containers (later described in this article).
To switch between Windows containers and Linux containers, right click the icon in the system tray and click the button to switch container operating system Doing this will stop the current containers that are running, and make them unaccessible until the container OS is switched back.
Additionally, if you have WSL or WSL2 installed on your desktop, you might want to install the Linux Kernel for Windows. Instructions can be found here. This requires the Windows Subsystem for Linux feature. This will allow for containers to be accessed by WSL operating systems, as well as the efficiency gain from running WSL operating systems in docker. It is also preferred to use Windows terminal for this.
Follow Microsoft's instructions that can be found here
If using the latest edge version of 2019, be prepared to only work in powershell, as it is only a servercore image (no desktop interface). When starting this machine, it will login and go straight to a powershell window. It is reccomended to install text editors and other tools using Chocolatey.
After installing, these commands will work:
# Display the version of docker installed:
docker version
# Pull, create, and run 'hello-world':
docker run hello-world
Windows Server 2016 is not able to run Linux images.
Windows Server Build 2004 is capable of running both linux and windows containers simultaneously through Hyper-V isolation. When running containers, use the --isolation=hyperv
command, which will isolate the container using a seperate kernel instance.
It is very important that you always know the current version of Docker you are currently running on at any point in time. This is very helpful because you get to know what features are compatible with what you have running. This is also important because you know what containers to run from the docker store when you are trying to get template containers. That said let see how to know which version of docker we have running currently.
docker version
shows which version of docker you have running.Get the server version:
$ docker version --format '{{.Server.Version}}'
1.8.0
You can also dump raw JSON data:
$ docker version --format '{{json .}}'
{"Client":{"Version":"1.8.0","ApiVersion":"1.20","GitCommit":"f5bae0a","GoVersion":"go1.4.2","Os":"linux","Arch":"am"}
Your basic isolated Docker process. Containers are to Virtual Machines as threads are to processes. Or you can think of them as chroots on steroids.
docker create
creates a container but does not start it.docker rename
allows the container to be renamed.docker run
creates and starts a container in one operation.docker rm
deletes a container.docker update
updates a container's resource limits.Normally if you run a container without options it will start and stop immediately, if you want keep it running you can use the command, docker run -td container_id
this will use the option -t
that will allocate a pseudo-TTY session and -d
that will detach automatically the container (run container in background and print container ID).
If you want a transient container, docker run --rm
will remove the container after it stops.
If you want to map a directory on the host to a docker container, docker run -v $HOSTDIR:$DOCKERDIR
. Also see Volumes.
If you want to remove also the volumes associated with the container, the deletion of the container must include the -v
switch like in docker rm -v
.
There's also a logging driver available for individual containers in docker 1.10. To run docker with a custom log driver (i.e., to syslog), use docker run --log-driver=syslog
.
Another useful option is docker run --name yourname docker_image
because when you specify the --name
inside the run command this will allow you to start and stop a container by calling it with the name the you specified when you created it.
docker start
starts a container so it is running.docker stop
stops a running container.docker restart
stops and starts a container.docker pause
pauses a running container, "freezing" it in place.docker unpause
will unpause a running container.docker wait
blocks until running container stops.docker kill
sends a SIGKILL to a running container.docker attach
will connect to a running container.If you want to detach from a running container, use Ctrl + p, Ctrl + q
.If you want to integrate a container with a host process manager, start the daemon with -r=false
then use docker start -a
.
If you want to expose container ports through the host, see the exposing ports section.
Restart policies on crashed docker instances are covered here.
You can limit CPU, either using a percentage of all CPUs, or by using specific cores.
For example, you can tell the cpu-shares
setting. The setting is a bit strange -- 1024 means 100% of the CPU, so if you want the container to take 50% of all CPU cores, you should specify 512. See https://goldmann.pl/blog/2014/09/11/resource-management-in-docker/#_cpu for more:
docker run -it -c 512 agileek/cpuset-test
You can also only use some CPU cores using cpuset-cpus
. See https://agileek.github.io/docker/2014/08/06/docker-cpuset/ for details and some nice videos:
docker run -it --cpuset-cpus=0,4,6 agileek/cpuset-test
Note that Docker can still see all of the CPUs inside the container -- it just isn't using all of them. See https://github.com/docker/docker/issues/20770 for more details.
You can also set memory constraints on Docker:
docker run -it -m 300M ubuntu:14.04 /bin/bash
Linux capabilities can be set by using cap-add
and cap-drop
. See https://docs.docker.com/engine/reference/run/#/runtime-privilege-and-linux-capabilities for details. This should be used for greater security.
To mount a FUSE based filesystem, you need to combine both --cap-add and --device:
docker run --rm -it --cap-add SYS_ADMIN --device /dev/fuse sshfs
Give access to a single device:
docker run -it --device=/dev/ttyUSB0 debian bash
Give access to all devices:
docker run -it --privileged -v /dev/bus/usb:/dev/bus/usb debian bash
More info about privileged containers here.
docker ps
shows running containers.docker logs
gets logs from container. (You can use a custom log driver, but logs is only available for json-file
and journald
in 1.10).docker inspect
looks at all the info on a container (including IP address).docker events
gets events from container.docker port
shows public facing port of container.docker top
shows running processes in container.docker stats
shows containers' resource usage statistics.docker diff
shows changed files in the container's FS.docker ps -a
shows running and stopped containers.
docker stats --all
shows a list of all containers, default shows just running.
docker cp
copies files or folders between a container and the local filesystem.docker export
turns container filesystem into tarball archive stream to STDOUT.docker exec
to execute a command in container.To enter a running container, attach a new shell process to a running container called foo, use: docker exec -it foo /bin/bash
.
Images are just templates for docker containers.
docker images
shows all images.docker import
creates an image from a tarball.docker build
creates image from Dockerfile.docker commit
creates image from a container, pausing it temporarily if it is running.docker rmi
removes an image.docker load
loads an image from a tar archive as STDIN, including images and tags (as of 0.7).docker save
saves an image to a tar archive stream to STDOUT with all parent layers, tags & versions (as of 0.7).docker history
shows history of image.docker tag
tags an image to a name (local or registry).While you can use the docker rmi
command to remove specific images, there's a tool called docker-gc that will safely clean up images that are no longer used by any containers. As of docker 1.13, docker image prune
is also available for removing unused images. See Prune.
Load an image from file:
docker load < my_image.tar.gz
Save an existing image:
docker save my_image:my_tag | gzip > my_image.tar.gz
Import a container as an image from file:
cat my_container.tar.gz | docker import - my_image:my_tag
Export an existing container:
docker export my_container | gzip > my_container.tar.gz
Loading an image using the load
command creates a new image including its history.
Importing a container as an image using the import
command creates a new image excluding the history which results in a smaller image size compared to loading an image.
Docker has a networks feature. Docker automatically creates 3 network interfaces when you install it (bridge, host none). A new container is launched into the bridge network by default. To enable communication between multiple containers, you can create a new network and launch containers in it. This enables containers to communicate to each other while being isolated from containers that are not connected to the network. Furthermore, it allows to map container names to their IP addresses. See working with networks for more details.
docker network create
NAME Create a new network (default type: bridge).docker network rm
NAME Remove one or more networks by name or identifier. No containers can be connected to the network when deleting it.docker network ls
List networksdocker network inspect
NAME Display detailed information on one or more networks.docker network connect
NETWORK CONTAINER Connect a container to a networkdocker network disconnect
NETWORK CONTAINER Disconnect a container from a networkYou can specify a specific IP address for a container:
# create a new bridge network with your subnet and gateway for your ip block
docker network create --subnet 203.0.113.0/24 --gateway 203.0.113.254 iptastic
# run a nginx container with a specific ip in that block
$ docker run --rm -it --net iptastic --ip 203.0.113.2 nginx
# curl the ip from any other place (assuming this is a public ip block duh)
$ curl 203.0.113.2
A repository is a hosted collection of tagged images that together create the file system for a container.
A registry is a host -- a server that stores repositories and provides an HTTP API for managing the uploading and downloading of repositories.
Docker.com hosts its own index to a central registry which contains a large number of repositories. Having said that, the central docker registry does not do a good job of verifying images and should be avoided if you're worried about security.
docker login
to login to a registry.docker logout
to logout from a registry.docker search
searches registry for image.docker pull
pulls an image from registry to local machine.docker push
pushes an image to the registry from local machine.You can run a local registry by using the docker distribution project and looking at the local deploy instructions.
Also see the mailing list.
The configuration file. Sets up a Docker container when you run docker build
on it. Vastly preferable to docker commit
.
Here are some common text editors and their syntax highlighting modules you could use to create Dockerfiles:
ADD
and use COPY
instead.--chown=<user>:<group>
to give ownership to another user/group. (Same for ADD
.)The versioned filesystem in Docker is based on layers. They're like git commits or changesets for filesystems.
Links are how Docker containers talk to each other through TCP/IP ports. Atlassian show worked examples. You can also resolve links by hostname.
This has been deprecated to some extent by user-defined networks.
NOTE: If you want containers to ONLY communicate with each other through links, start the docker daemon with -icc=false
to disable inter process communication.
If you have a container with the name CONTAINER (specified by docker run --name CONTAINER
) and in the Dockerfile, it has an exposed port:
EXPOSE 1337
Then if we create another container called LINKED like so:
docker run -d --link CONTAINER:ALIAS --name LINKED user/wordpress
Then the exposed ports and aliases of CONTAINER will show up in LINKED with the following environment variables:
$ALIAS_PORT_1337_TCP_PORT
$ALIAS_PORT_1337_TCP_ADDR
And you can connect to it that way.
To delete links, use docker rm --link
.
Generally, linking between docker services is a subset of "service discovery", a big problem if you're planning to use Docker at scale in production. Please read The Docker Ecosystem: Service Discovery and Distributed Configuration Stores for more info.
Docker volumes are free-floating filesystems. They don't have to be connected to a particular container. You can use volumes mounted from data-only containers for portability. As of Docker 1.9.0, Docker has named volumes which replace data-only containers. Consider using named volumes to implement it rather than data containers.
Volumes are useful in situations where you can't use links (which are TCP/IP only). For instance, if you need to have two docker instances communicate by leaving stuff on the filesystem.
You can mount them in several docker containers at once, using docker run --volumes-from
.
Because volumes are isolated filesystems, they are often used to store state from computations between transient containers. That is, you can have a stateless and transient container run from a recipe, blow it away, and then have a second instance of the transient container pick up from where the last one left off.
See advanced volumes for more details. Container42 is also helpful.
You can map MacOS host directories as docker volumes:
docker run -v /Users/wsargent/myapp/src:/src
You can use remote NFS volumes if you're feeling brave.
You may also consider running data-only containers as described here to provide some data portability.
Be aware that you can mount files as volumes.
Exposing incoming ports through the host container is fiddly but doable.
This is done by mapping the container port to the host port (only using localhost interface) using -p
:
docker run -p 127.0.0.1:$HOSTPORT:$CONTAINERPORT \
--name CONTAINER \
-t someimage
You can tell Docker that the container listens on the specified network ports at runtime by using EXPOSE:
EXPOSE <CONTAINERPORT>
Note that EXPOSE
does not expose the port itself - only -p
will do that.
To expose the container's port on your localhost's port, run:
iptables -t nat -A DOCKER -p tcp --dport <LOCALHOSTPORT> -j DNAT --to-destination <CONTAINERIP>:<PORT>
If you're running Docker in Virtualbox, you then need to forward the port there as well, using forwarded_port. Define a range of ports in your Vagrantfile like this so you can dynamically map them:
Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
...
(49000..49900).each do |port|
config.vm.network :forwarded_port, :host => port, :guest => port
end
...
end
If you forget what you mapped the port to on the host container, use docker port
to show it:
docker port CONTAINER $CONTAINERPORT
This is where general Docker best practices and war stories go:
Compose is a tool for defining and running multi-container Docker applications. With Compose, you use a YAML file to configure your application’s services. Then, with a single command, you create and start all the services from your configuration. To learn more about all the features of Compose, see the list of features.
By using the following command you can start up your application:
docker-compose -f <docker-compose-file> up
You can also run docker-compose in detached mode using -d flag, then you can stop it whenever needed by the following command:
docker-compose stop
You can bring everything down, removing the containers entirely, with the down command. Pass --volumes
to also remove the data volume.
This is where security tips about Docker go. The Docker security page goes into more detail.
First things first: Docker runs as root. If you are in the docker
group, you effectively have root access. If you expose the docker unix socket to a container, you are giving the container root access to the host.
Docker should not be your only defense. You should secure and harden it.
For an understanding of what containers leave exposed, you should read Understanding and Hardening Linux Containers by Aaron Grattafiori. This is a complete and comprehensive guide to the issues involved with containers, with a plethora of links and footnotes leading on to yet more useful content. The security tips following are useful if you've already hardened containers in the past, but are not a substitute for understanding.
For greatest security, you want to run Docker inside a virtual machine. This is straight from the Docker Security Team Lead -- slides / notes. Then, run with AppArmor / seccomp / SELinux / grsec etc to limit the container permissions. See the Docker 1.10 security features for more details.
Docker image ids are sensitive information and should not be exposed to the outside world. Treat them like passwords.
See the Docker Security Cheat Sheet by Thomas Sjögren: some good stuff about container hardening in there.
Check out the docker bench security script, download the white papers.
Snyk's 10 Docker Image Security Best Practices cheat sheet
You should start off by using a kernel with unstable patches for grsecurity / pax compiled in, such as Alpine Linux. If you are using grsecurity in production, you should spring for commercial support for the stable patches, same as you would do for RedHat. It's $200 a month, which is nothing to your devops budget.
Since docker 1.11 you can easily limit the number of active processes running inside a container to prevent fork bombs. This requires a linux kernel >= 4.3 with CGROUP_PIDS=y to be in the kernel configuration.
docker run --pids-limit=64
Also available since docker 1.11 is the ability to prevent processes from gaining new privileges. This feature have been in the linux kernel since version 3.5. You can read more about it in this blog post.
docker run --security-opt=no-new-privileges
From the Docker Security Cheat Sheet (it's in PDF which makes it hard to use, so copying below) by Container Solutions:
Turn off interprocess communication with:
docker -d --icc=false --iptables
Set the container to be read-only:
docker run --read-only
Verify images with a hashsum:
docker pull debian@sha256:a25306f3850e1bd44541976aa7b5fd0a29be
Set volumes to be read only:
docker run -v $(pwd)/secrets:/secrets:ro debian
Define and run a user in your Dockerfile so you don't run as root inside the container:
RUN groupadd -r user && useradd -r -g user user
USER user
There's also work on user namespaces -- it is in 1.10 but is not enabled by default.
To enable user namespaces ("remap the userns") in Ubuntu 15.10, follow the blog example.
The Docker roadmap talks about seccomp support.There is an AppArmor policy generator called bane, and they're working on security profiles.
Sources:
The new Data Management Commands have landed as of Docker 1.13:
docker system prune
docker volume prune
docker network prune
docker container prune
docker image prune
docker system df
presents a summary of the space currently used by different docker objects.
docker build -t htop - << EOF
FROM alpine
RUN apk --no-cache add htop
EOF
alias dl='docker ps -l -q'
docker run ubuntu echo hello world
docker commit $(dl) helloworld
docker commit -run='{"Cmd":["postgres", "-too -many -opts"]}' $(dl) postgres
docker inspect $(dl) | grep -wm1 IPAddress | cut -d '"' -f 4
Or with jq installed:
docker inspect $(dl) | jq -r '.[0].NetworkSettings.IPAddress'
Or using a go template:
docker inspect -f '{{ .NetworkSettings.IPAddress }}' <container_name>
Or when building an image from Dockerfile, when you want to pass in a build argument:
DOCKER_HOST_IP=`ifconfig | grep -E "([0-9]{1,3}\.){3}[0-9]{1,3}" | grep -v 127.0.0.1 | awk '{ print $2 }' | cut -f2 -d: | head -n1`
echo DOCKER_HOST_IP = $DOCKER_HOST_IP
docker build \
--build-arg ARTIFACTORY_ADDRESS=$DOCKER_HOST_IP
-t sometag \
some-directory/
docker inspect -f '{{range $p, $conf := .NetworkSettings.Ports}} {{$p}} -> {{(index $conf 0).HostPort}} {{end}}' <containername>
for i in $(docker ps -a | grep "REGEXP_PATTERN" | cut -f1 -d" "); do echo $i; done
docker run --rm ubuntu env
docker kill $(docker ps -q)
docker rm -f $(docker ps -qa)
docker ps -a | grep 'weeks ago' | awk '{print $1}' | xargs docker rm
docker rm -v $(docker ps -a -q -f status=exited)
docker stop $(docker ps -aq) && docker rm -v $(docker ps -aq)
docker rmi $(docker images -q -f dangling=true)
docker rmi $(docker images -q)
As of Docker 1.9:
docker volume rm $(docker volume ls -q -f dangling=true)
In 1.9.0, the filter dangling=false
does not work - it is ignored and will list all volumes.
docker images -viz | dot -Tpng -o docker.png
RUN
layer - This should be done in the same layer as other apt
commands. Otherwise, the previous layers still persist the original information and your images will still be fat.
RUN {apt commands} \
&& apt-get clean \
&& rm -rf /var/lib/apt/lists/* /tmp/* /var/tmp/*
ID=$(docker run -d image-name /bin/bash)
docker export $ID | docker import – flat-image-name
ID=$(docker run -d image-name /bin/bash)
(docker export $ID | gzip -c > image.tgz)
gzip -dc image.tgz | docker import - flat-image-name
To check the CPU, memory, and network I/O usage of a single container, you can use:
docker stats <container>
For all containers listed by ID:
docker stats $(docker ps -q)
For all containers listed by name:
docker stats $(docker ps --format '{{.Names}}')
For all containers listed by image:
docker ps -a -f ancestor=ubuntu
Remove all untagged images:
docker rmi $(docker images | grep “^” | awk '{split($0,a," "); print a[3]}')
Remove container by a regular expression:
docker ps -a | grep wildfly | awk '{print $1}' | xargs docker rm -f
Remove all exited containers:
docker rm -f $(docker ps -a | grep Exit | awk '{ print $1 }')
Be aware that you can mount files as volumes. For example you can inject a configuration file like this:
# copy file from container
docker run --rm httpd cat /usr/local/apache2/conf/httpd.conf > httpd.conf
# edit file
vim httpd.conf
# start container with modified configuration
docker run --rm -it -v "$PWD/httpd.conf:/usr/local/apache2/conf/httpd.conf:ro" -p "80:80" httpd
Here's how to contribute to this cheat sheet.
Click README.md <-- this link
TensorFlow Docker requirements Install Docker on your local host machine. For GPU support on Linux, install nvidia-docker. Note: To run the docker command without sudo, create the docker group and add
介绍 (Introduction) Server automation now plays an essential role in systems administration, due to the disposable nature of modern application environments. Configuration management tools such as Ansib
cheatsheet 复制到container (不用区分文件还是文件夹) docker cp 文件路径 container_id:目录路径 Login docker login -u user -p passwd Pull docker pull image:tag Push(push之前要先登录login) docker push image/tag Run docker run -it --
安装环境 操作系统:Ubuntu Xenial 16.04 (LTS) Go语言版本:1.9.4 配置Ubuntu更新源(清华大学) $ sudo vim /etc/apt/sources.list # deb cdrom:[Ubuntu 16.04 LTS _Xenial Xerus_ - Release amd64 (20160420.1)]/ xenial main restricted d
目录 1. 创建对象 2. 显示和查找资源 3. 更新资源 4. 修补资源 5. 编辑资源 6. Scale 资源 7. 删除资源 8. 与运行中的 Pod 交互 9. 与节点和集群交互 10. 资源类型 11. 格式化输出 12. Kubectl 详细输出与调试 1. 创建对象 $ kubectl create -f ./my-manifest.yaml # 创建资源
? 节选自 Awesome CheatSheet/Docker CheatSheet,对来自官方文档及 Docker Links 中链接内容的归档整理,包含了日常工作中常用的 Docker 概念与命令,如果对于 Linux 常用操作尚不熟悉的可以参考 Linux Commands CheatSheet。 Docker CheatSheet | Docker 配置与实践清单 Docker 是一个开源
1.docker镜像优化的配置 2 避免泄露构建密钥 构建密钥是只在构建 Docker 镜像时需要的凭证(不是在运行时)。例如,你可能想在你的镜像中包含某个应用程序的一个编译版本,这个应用的源代码是闭源的,并且其 Git 存储库是有访问保护的。在构建镜像时,你需要克隆 Git 存储库(这需要构建密钥,例如该存储库的 SSH 访问密钥),从源代码构建应用程序,然后再删除源代码(和密钥)。 “泄露“
容器编排技术 -- kubectl Cheat Sheet 1 Kubectl 自动补全 2 Kubectl 上下文和配置 3 创建对象 4 显示和查找资源 5 更新资源 6 修补资源 7 编辑资源 8 Scale资源 9 删除资源 10 与运行中的 Pod 交互 11 与节点和集群交互 12 资源类型 12.1 格式化输出 12.2 Kubectl 详细输出与调试 更多内容请参考: Kube
ROS:http://wiki.ros.org/ 参考wiki.ros.org上的 Ubuntu install of ROS Kinetic 以及之前的博客:Docker尝试,在Docker中配置ROS Docker GUI Tutorial # 运行容器 GUI 且使用Nidia显卡 xhost+ sudo docker run --ipc=host --runtime=nvidia -e $
文章转载自: https://learnku.com/articles/22822 前面两篇文章,基于Docker搭建的MySQL主从复制。后来有网友QQ问我有没有Docker的相关文章,说没有接触过这个Docker,不知道如何上手,今天就借花献佛,转载一篇比较全面的Docker大全文章。本篇篇幅比较长,请耐心看完。如果转载请注明原文出处。文章来自:learnku编程者知识社区,一个很不错的社区。
Docker containers opened a world of possibilities for the tech community, hassles in setting up new software were decreased unlike old times when a mess was to be sorted by a grievous format, it reduc
bitnami如何使用 I’ve been using it for around 2 years now to build out custom workflow interfaces, like those used for Laboratory Information Management Systems (LIMs), Computer Vision pre and postprocess
CHEAT 是一个微型的 C 语言单元测试框架。没有任何依赖和安装配置,使用简单只需一个头文件和一个测试用例即可,示例代码: #include <cheat.h>CHEAT_TEST(mathematics_still_work, cheat_assert(2 + 2 == 4); cheat_assert_not(2 + 2 == 5);)
Cheat Engine 是一款开源的内存修改工具 ,它允许你修改游戏或软件内存数据,以达到各种非常规目的。 它包括16进制编辑,反汇编程序,内存查找工具。 它功能丰富专业,与同类修改工具相比,更像是一个Crack工具。 它具有强大的反汇编功能,且自身附带了外挂制作工具,甚至可用之直接生成修改器。 它具有良好的64位支持。支持插件,在官方网站可找到大量现成的游戏 Cheat Tables。
Cheat Sheets A searchable site of hints and tips! Image Credit ccbyplz A place for all my cheat sheets to live! This is a list of stuff that I have put down that I continuallyreference. Current list o
Linux 命令行的强大在于其灵活及多样化,各个Linux命令都带有它自己专属的命令行选项和参数。混合并匹配这些命令,甚至还可以通过管道和重 定向来联结不同的命令。理论上讲,你可以借助几个基本的命令来产生数以百计的使用案例。甚至对于浸淫多年的管理员而言,也难以完全使用它们。那正是命令行小抄成为我们救命稻草的一刻。 我知道联机手册页(man)仍然是我们的良师益友,但我们想通过我们能自行支配的快速参考
Concise Cheat Sheets This repository is a collection of Cheat Sheets for programming languages andtools. Currently: A Haskell Cheat Sheet -- version 1.1 A Python Cheat Sheet -- version 0.1 (work in pr
emoji-cheat-sheet This cheat sheet is automatically generated from GitHub Emoji API and Unicode Full Emoji List. Table of Contents Smileys & Emotion People & Body Animals & Nature Food & Drink Travel