Make Your Own Private Bitcoin Node to Anonymize Transactions over Tor

This article has been updated to include instructions to update Tor to the latest version, as I discovered it was using a really old version in the repositories. There was also an addition to tell the systemd services to safely shutdown bitcoin when it is stopped. There has also been an amendment to the bitcoin.conf file I made. You can find the amendments in the Tor, Installing Bitcoin Core, and Autostart sections

Chances are you’ve heard of Bitcoin, the anonymous and secure cryptocurrency which has made waves over many years. One of the main issues I see is that people are trusting others to handle their transactions. So I set about purchasing a tiny Dell netbook with a measly Intel Atom CPU, 2GB of RAM, and a 240GB SSD to act as my primary wallet for cryptocurrency, and is more or less my bank. The laptop has a fully encrypted drive, and I back up the keys for my wallet and have them in three different places. However when you are running a full version of the core wallets, that means you have to store a whole copy of the entire block chain on the device using it. Currently the Bitcoin blockchain is around 200GB if I recall, and that’s a lot of data to hold onto just to transact. Not to mention the whole idea of my netbook was to only be on when I needed to transact as it’s most secure when it is powered off. So obviously running the blockchain on the laptop was not the most ideal option as it would always have to be on. Not to mention I want to further anonymize the connections coming in and out, so I wanted to tunnel all of the traffic for the node over a VPN such as Private Internet Access, with a VPN killswitch so if the VPN doesn’t work it won’t connect, as well as bolt The Onion Relay (TOR) on top of it to further anonymize the entire transactions. The advantage of all this will be to allow any device on my LAN to transact with the blockchain network directly using my node to send and receive my transactions rather than trusting other people. The other advantage this has is since I am running a full copy of the blockchain I am also helping support the Bitcoin network by providing another peer with a full copy of the blockchain. However this guide will not cover the VPN aspect, but we will cover how to bolt on Tor as well as build your own node entirely.

I originally got the idea from where they have a lot of these projects, and I am already running a Monero node on a Pi 3B+ with a 128GB flash drive. But for this tutorial I am merging the ideas from pinode, along with the Thundroid tutorial, and adding some of my own twists and spins. I chose an Odroid Home Cloud 2 as it allows for a native SATA hard drive, and the HC2 variant allows for 3.5″ disks where as the HC1 allows for 2.5″ disks. Either version is fine, and if you really wanted you could probably go for the straight XU4 version, or even a Raspberry Pi 3B+ if you use a 512GB flash drive or larger, or a kit to allow additional drives. But for the sake of this tutorial we will be discussing the Odroid platform, however you can use whatever platform you like. Technically if you wanted to, you could use a full dedicated PC, but that seems like a waste of hardware and will be far less power efficient. I prefer the Odroid over the Raspberry Pi as it’s a more powerful hardware platform.

Hardware (all links are to Ameri Droid as I am in the USA):

Odroid Home Cloud 2 with RTC battery, 1TB Seagate Iron Wolf HDD, and Model 3 Wifi NIC.

The reason I specified using a NAS drive is that this drive will be on 24/7 and always writing as well as reading data. NAS drives are specifically optimized for this kind of behavior, and will therefore be more reliable. You can use a non-NAS drive just fine, but in the long term a NAS drive is best.

Odroid HC2 pictured in the clear acryllic case option with the wireless NIC inserted, and 16GB flash card before it was flashed with the Ubuntu 18.04.1 OS

Optional Hardware:

Odroid HC2 being worked on at my desk with the UART connector kit pictured.

First things first we have to connect it to the internet. So if you are planning on using wifi please follow the wiki here for nmcli for the Odroid. If using the UART console connection follow this tutorial here. You will need to flash Ubuntu 18.04 minimal image which can be downloaded here and then use Etcher to flash it to a MicroSD. Once that is down put it in the Odroid, and boot it up and either SSH or connect via console to it. Either way the credentials on start are:
username: root
password: odroid

For Raspberry Pi users you will have to look up the credentials for the image you are using.


We’ll need to take care of some things first before we actually make it a Bitcoin node. So first let’s create a new user with a secure password and superuser rights and change the root password. Don’t forget to change “USER” to what you want.

[email protected]:~# passwd
[email protected]:~# adduser USER
[email protected]:~# usermod -aG sudo USER
[email protected]:~# adduser bitcoin

Now we need to update the system and change the timezone and locale data, as well as change the hostname in both /etc/hosts and /etc/hostname to match. I named mine “btcdroid” but you can make it whatever you want:

[email protected]:~# apt update
[email protected]:~# apt dist-upgrade -y
[email protected]:~# apt install htop git curl bash-completion jq
[email protected]:~# dpkg-reconfigure tzdata
[email protected]:~# dpkg-reconfigure locales
[email protected]:~# nano /etc/hosts
[email protected]:~# nano /etc/hostname

Mount the Hard Drive

Now we need to mount the hard drive. In my case the hard drive was brand new and unformatted, so I had to do that first, you can follow the instructions here at Digital Ocean if you are in the same situation. Regardless, once you have a formatted drive compatible with Linux we can proceed.

We will need to get the UUID of the partition that has been created. To do that is very simple we run the lsblk command and it will spit out the names and UUID of all drives.

[email protected]:~# lsblk --fs

After running that command you should see something like this. We will need to notate the UUID it has given us for the next steps.

Now we need to edit the fstab with nano and add a whole new line. Replace 123456 with the UUID given from the command above

[email protected]:~# nano /etc/fstab
# New Line in /etc/fstab
UUID=123456 /mnt/hdd ext4 noexec,defaults 0 0 

Awesome now the fstab has been modified, now we need to create the mount point, mount it, check it, and set the owner.

[email protected]:~# mkdir /mnt/hdd
[email protected]:~# mount -a
[email protected]:~# df /mnt/hdd 

At this point if everything was done correctly you should see something similar to this.

Now let’s give permissions to that entire hard drive to the bitcoin user we made earlier

[email protected]:~# chown -R bitcoin:bitcoin /mnt/hdd/ 

Moving Swap to the HDD

Now we need to move the swap file to the HDD. So we need to install a package and then do some configuration changes.

[email protected]:~# apt install dphys-swapfile
[email protected]:~# nano /etc/dphys-swapfile
#Add the following lines
[email protected]:~# dphys-swapfile setup
[email protected]:~# dphys-swapfile swapon
[email protected]:~# shutdown -r now 

Hardening The Security

Now it should be reconfigured to have a 2GB swap file on the hard drive, and should be rebooting. At this point log back in as the regular user and not as root. Because we are about to disable root via SSH, however if you are using the optional UART serial connection kit, you can still login as root that way. Now let’s continue on and remove the old swap file.

SSH Hardening

We need to lock down remote access to SSH, and Digital Ocean has a great guide going over SSH security. I highly recommend disabling password logins and requiring an SSH key pair to be generated. You can read the tutorial here, but we will definitely need to be disabling root access as well. It’s a major security risk if root is allowed, as everyone knows Linux has a root user.

Type the following command to edit the sshd_config file.

[email protected]:~$ sudo nano /etc/ssh/sshd_config

#Find the following line PermitRootLogin yes #Change it to no so it looks like below PermitRootLogin no #Save and quit
[email protected]:~$ sudo service sshd restart

That will disable root login, but again I highly recommend making it only allow logins with SSH key pairs as it is far more secure than a password.


So one of my favorite tools, which I have written about before is uncomplicated firewall. We are going to allow only pinholes for the firewall to allow communication through as well as limit ssh connections so it is less likely to be brute forced. We will also be adding some defenses for brute forcing in a bit.

The line ufw allow from, below assumes that the IP address of your btcdroid is something like, the xxx being any number from 0 to 255. If your IP address is, you must adapt this line to ufw allow from Otherwise you will lock yourself out for good unless you connect the UART serial connection kit.

[email protected]:~$ sudo apt install ufw
[email protected]:~$ sudo ufw default deny incoming
[email protected]:~$ sudo ufw default allow outgoing

# make sure to use the correct subnet mask and IP ranges. (see warning above)
[email protected]:~$ sudo ufw allow from to any port 22 comment 'allow SSH from local LAN'
[email protected]:~$ sudo ufw allow 9735 comment 'allow Lightning'
[email protected]:~$ sudo ufw allow 8333 comment 'allow Bitcoin mainnet'
[email protected]:~$ sudo ufw allow 18333 comment 'allow Bitcoin testnet'
[email protected]:~$ sudo ufw enable
[email protected]:~$ sudo systemctl enable ufw
[email protected]:~$ sudo ufw status

Now we should install Fail2Ban, which I have talked about often. This will make it so after five unsuccessful attempts at SSH it blocks the IP for ten minutes. Making a brute force almost impossible to conduct.

[email protected]:~$ sudo apt install fail2ban

Increase open file limit

In case your BTCDroid is swamped with internet requests (honest or malicious due to a DDoS attack), you will quickly encounter the can't accept connection: too many open files error. This is due to a limit on open files (representing individual tcp connections) that is set too low.

Edit the following three files, add the additional line(s) right before the end comment, save and exit.

[email protected]:~$ sudo nano /etc/security/limits.conf
#add/change the following lines
*    soft nofile 128000
*    hard nofile 128000
root soft nofile 128000
root hard nofile 128000

[email protected]:~$ sudo nano /etc/pam.d/common-session #add the following session required
[email protected]:~$ sudo nano /etc/pam.d/common-session-noninteractive #add the following session required

Installing Bitcoin Core

We’re finally ready to start with the fun parts. These parts were mostly derived from, but seem to work perfectly fine for the Odroid HC2, albeit with some tweaks we have already performed specific to the Odroid platform.

First we need to install our dependencies:

[email protected]:~$ sudo apt install autoconf libevent-dev libtool libssl-dev libboost-all-dev libminiupnpc-dev -y 

Now we need to make a directory to download our files into, and ultimately download those files using git

[email protected]:~$ mkdir ~/bin
[email protected]:~$ cd ~/bin
[email protected]:~$ git clone -b 0.17

Now after it’s downloaded we are going to configure, compile, and install the files. Now I tell it in the final command to run six jobs at the same time since the Odroid has eight cores so it can run faster. You may want to reduce that number to two with a Raspberry Pi. You can also run it without the “-jX” switch to just run as a single job, although that may take a couple hours. Once you run the make command, go make dinner or something because this will take an hour or two even on the Odroid XU4’s eight core Samsung Exynos 5422 CPU.

[email protected]:~$ cd bitcoin
[email protected]:~$ ./
[email protected]:~$ ./configure --enable-upnp-default --disable-wallet
[email protected]:~$ make -j6
[email protected]:~$ sudo make install

Now we need to prepare the Bitcoin directory, we’re going to switch into the non super user we created earlier which we named bitcoin, although you can name it whatever you want. The most important thing is that this user only have permissions to administrate the bitcoin node itself and not able to make any system changes. This is the great thing about Linux in regards to security and permissions versus Windows. This in theory should isolate an attack from at work messing with just the bitcoin systems and not the system itself.

We use the Bitcoin daemon, called “bitcoind”, that runs in the background without user interface and stores all data in a the directory /home/bitcoin/.bitcoin. Instead of creating a real directory, we create a link that points to a directory on the external hard disk.

[email protected]:~$ sudo su bitcoin

# add symbolic link that points to the external hard drive
[email protected]:~$ mkdir /mnt/hdd/bitcoin
[email protected]:~$ ln -s /mnt/hdd/bitcoin /home/bitcoin/.bitcoin

# Navigate to home directory and check the symbolic link (the target must not be red). 
[email protected]:~$ cd ~
[email protected]:~$ ls -la

Now we need to configure the Bitcoin daemon, and make sure to set an extremely secure password and username seperate from your username and password on the system, and then we will log out of the bitcoin user to setup Tor.

[email protected]:~$ nano /home/bitcoin/.bitcoin/bitcoin.conf

# BTCDroid: bitcoind configuration
# /home/bitcoin/.bitcoin/bitcoin.conf

# Bitcoind options

# Connection settings

# Optimizations for Odroid Hardware

#Optimizations for Raspberry Pi 3B.
#I commented out the ones for the ones I recommend for a Raspberry Pi 3B, just uncomment those, and comment out the Odroid ones for it to work

[email protected]:~$ exit

Tor IT Up

Now we get to install Tor to encapsulate all the traffic and encrypt as well as anonymize everything. So we are going to install Tor, but also add a repository to give us the most up to date Tor version, as the one in the default repositories is really old.

First we will be adding a couple entries to /etc/apt/sources.list.d/, add the GPG key to accept it, and then update our repository, and finally install Tor.

[email protected]:~$ sudo nano /etc/apt/sources.list.d/tor.list
#Add the following lines and then save and close
deb bionic main
deb-src bionic main
#save and exit
[email protected]:~$ curl | gpg --import
[email protected]:~$ gpg --export A3C4F0F979CAA22CDBA8F512EE8CBC9E886DDD89 | sudo apt-key add -
[email protected]:~$  apt update
[email protected]:~$  apt install tor tor-arm nyx

Now we need to configure Tor

[email protected]:~$ sudo nano /etc/tor/torrc
#add these settings to the bottom of the file
ControlPort 9051
CookieAuthentication 1
CookieAuthFileGroupReadable 1
HiddenServiceDir /var/lib/tor/bitcoin-service/
HiddenServicePort 8333
#save and exit
[email protected]:~$ sudo systemctl restart tor.service
#Get your Tor hostname
[email protected]:~$ sudo cat /var/lib/tor/bitcoin-service/hostname

That hostname it spits out with a “.onion” address, we’re going to need that, so notate what it just gave us with the last command.

configure everything to autostart

Now we need to program everything to start on boot, so we will make a SystemD service that will start our Bitcoin node how we would like it to start with it running as the bitcoin user and passing it through to Tor. Now I will give you the option at this point to either have it run only on Tor, or to allow it to run over Tor, IPv4, and IPv6. The Tor only option is more anonymous, but the other mode is like a dual mode so if Tor is down it can still sync, but it also will sync faster. The choice is yours, just remove the comment for the one you want where it says ExecStart and don’t forget to insert your hostname.onion we pulled from earlier where it asks for it in the ExecStart command. After that we will reboot and see if everything works. Make sure to put your username we created earlier on where it says USER_NAME

[email protected]:~$ sudo nano /etc/systemd/system/bitcoind.service

# BTCdroid systemd unit for bitcoind
# /etc/systemd/system/bitcoind.service

Description=Bitcoin daemon

#Uncomment the ExecStart string below to force the node to only run over Tor
#ExecStart= /usr/local/bin/bitcoind -datadir=/home/bitcoin/.bitcoin/data -daemon -proxy= -externalip=HOSTNAME.onion -conf=/home/bitcoin/.bitcoin/bitcoin.conf -listen -bind= -pid=/run/bitcoind/

#Uncomment the ExecStart string below to allow Tor, IPv4, and IPv6 connections
#ExecStart= /usr/local/bin/bitcoind -datadir=/home/bitcoin/.bitcoin/data -daemon -proxy= -externalip=HOSTNAME.onion -conf=/home/bitcoin/.bitcoin/bitcoin.conf -listen -discover -pid=/run/bitcoind/

#Tells Bitcoin to shutdown safely when stopped. 
ExecStop= /usr/local/bin/bitcoin-cli stop

# Creates /run/bitcoind owned by bitcoin

# Hardening measures

# Provide a private /tmp and /var/tmp.

# Mount /usr, /boot/ and /etc read-only for the process.

# Disallow the process and all of its children to gain
# new privileges through execve().

# Use a new /dev namespace only populated with API pseudo devices
# such as /dev/null, /dev/zero and /dev/random.

# Deny the creation of writable and executable memory mappings.

#save and exit
[email protected]:~$ sudo systemctl enable bitcoind.service
[email protected]:~$ sudo shutdown -r now 
[email protected]:~$ mkdir /home/USER_NAME/.bitcoin
[email protected]:~$ sudo cp /home/bitcoin/.bitcoin/bitcoin.conf /home/USER_NAME/.bitcoin/
[email protected]:~$ sudo chown USER_NAME:USER_NAME /home/USER_NAME/.bitcoin/bitcoin.conf

Now it should be restarting so give it a minute and reconnect as the user we created in the beginning. It may take a few minutes for the node to get its first connections, and then it will start pulling in the blocks. You can check the status with the bitcoin-cli command.

[email protected]:~$ bitcoin-cli getblockchaininfo 

It should display something like this, and as long as the number of blocks is increasing every few minutes, it is running fine. Bare in mind this could take a few days as we need to download at least 200GB at the time of writing to be up to date with the block chain.

Output of bitcoin-cli

In addition to checking the status of the blockchain download, you can monitor the traffic over Tor with Nyx.

[email protected]:~$ sudo nyx 
Seeing the traffic via Tor on Nyx

Auto Update Security Patches

Since this is a device we are going to leave on and unattended most likely. It’s best we have it auto apply any of the security related patches that may be out there so it can maintain itself. So let’s enable the unattended-upgrades package and configure it. The first step brings up an interactive prompt, and then we proceed to editing the files.

[email protected]:~$ sudo dpkg-reconfigure --priority=low unattended-upgrades
[email protected]:~$ sudo nano /etc/apt/apt.conf.d/50unattended-upgrades
#modify these lines in the file to look like the following, although you can make it reboot whenever you want. Make sure there is a semicolon at the end of each line. You can uncomment the "${distro_id}:${distro_codename}-updates"; line if you want it to update non security related packages too

#near the top of the file
Unattended-Upgrade::Allowed-Origins {
        // Extended Security Maintenance; doesn't necessarily exist for
        // every release and this system may not have it installed, but if
        // available, the policy for updates is such that unattended-upgrades
        // should also install from here by default.
//      "${distro_id}:${distro_codename}-updates";
//      "${distro_id}:${distro_codename}-proposed";
//      "${distro_id}:${distro_codename}-backports";

#below are spread out in the same file
Unattended-Upgrade::Remove-Unused-Kernel-Packages "true";
Unattended-Upgrade::Remove-Unused-Dependencies "true";
Unattended-Upgrade::Automatic-Reboot "true";
Unattended-Upgrade::Automatic-Reboot-Time "02:30";
#save and exit 

That’s it, you’re all finished. Let me know what you think or if you have any improvements to the project. I may eventually be hosting these on a Supermicro server in my rack with a ZFS array next year.

Make a super fast and lightweight WordPress on Ubuntu 18.04 with PHP 7.2, Nginx, and MariaDB

I’ve been building servers for a long while based on the ideas I learned a few years ago from, however I wanted to move on PHP 7.2 and I also wanted to begin a server migration project to have, of which I am the CTO, also host our newest members and since they really liked the speed of our WordPress stack. This is a wordpress stack we will build based on Ubuntu 18.04, Nginx, MariaDB, and PHP 7.2. We will even cover setting up lets encrypt. Just a note that I use Mailgun to deliver the emails, it’s free for up to 10,000 emails per month, and they have an easy to use WordPress plugin that makes it super easy to configure. There is of course far more that you can do to secure your server, and while we aren’t going to cover hosting multiple sites in this tutorial, you can understand how I made such a robust server stack on a Digital Ocean Virtual Private Server. You can use this referral code ( get you a few dollars free with Digital Ocean when you sign up, and it also helps support the costs of my own hosting.

So first things first go to Digital Ocean, which we use, or any other VPS or dedicated server provider you use and get the OS setup and do the basics so we can even login to the box, and point your domain at your server. Once that is done I like to start setting up security, disabling root, and allowing a username to have sudo rights.

For those new to Linux administration you can use these tutorials as to how to add new sudo users and setup ssh keys for for even more security but once that is done let’s move on to the basic security I use.

We definitely want the firewall on our box, but IPtables can be a pain to manage. So let’s begin installing things on the box for security. Including Uncomplicated Firewall which can easily manage firewall rules for us.

Install UFW, Fail2Ban, Nginx, and MariaDB:

In order to use a WordPress plugin for purging the NGINX cache that I talk about below, you have to install a custom version of NGINX. MariaDB is a drop-in replacement for MySQL. You can read about why people think it’s better, but from what I have mostly noticed is that it is incredibly fast compared to MySQL. The MariaDB website has a convenient tool for configuring the correct repositories in your Ubuntu distro. From the command line:

sudo apt update 
sudo apt dist-upgrade -y 
sudo apt install ufw fail2ban
sudo apt-key adv --keyserver --recv-keys 3050AC3CD2AE6F03
sudo sh -c "echo 'deb /' >> /etc/apt/sources.list.d/nginx.list"
sudo apt update
sudo apt install nginx-custom
sudo ufw limit ssh
sudo ufw allow 'Nginx Full'
sudo ufw enable
sudo apt-key adv --recv-keys --keyserver hkp:// 0xF1656F24C74CD1D8 sudo add-apt-repository 'deb [arch=amd64,arm64,ppc64el] bionic main' 
sudo apt install mariadb-server

When the following screen comes up, make sure you provide a good secure password that is different from the password you used for your user account.

Next, lock down your MariaDB instance by running:

sudo mysql_secure_installation

Since you’ve already set up a secure password for your root user, you can safely answer “no” to the question asking you to create a new root password. Answer “Yes” to all of the other questions. Now we can set up a separate MariaDB account and database for our WordPress instance. At the command prompt type the following:

sudo mysql -u root -p

Type in your password when prompted. This will open up a MariaDB shell session. Everything you type here is treated as a SQL query, so make sure you end every line with a semicolon! This is very easy to forget. Here are the commands you need to type in to create a new database, user, and assign privileges to that user:

MariaDB [(none)]> CREATE DATABASE mywpdb DEFAULT CHARACTER SET utf8 COLLATE utf8_unicode_ci;
MariaDB [(none)]> GRANT ALL ON mywpdb.* TO 'mywpdbuser'@'localhost' IDENTIFIED BY 'securepassword';
MariaDB [(none)]> quit

Note that although it’s customary to use ALL CAPS to write SQL statements like this, it is not strictly necessary. Also, where I’ve used mywpdb, mywpdbuser, and securepassword make sure to put your own choices. The last thing you want is someone knowing you had an easy to guess database name and password.

Fail2Ban Installation and Setup:

Fail2Ban is an intrusion prevention software framework that protects servers from brute-force attacks. It’s probably one of my all time favorite security tools as it’s very robust and flexible. In order to make modifications to Fail2Ban we need to make a local copy that we can modify so we can preserve changes.

sudo cp /etc/fail2ban/jail.conf /etc/fail2ban/jail.local

Now open the newly made file so we can edit it

sudo nano /etc/fail2ban/jail.local

I recommend reading this guide from Digital Ocean on Fail2Ban with Nginx and follow the tutorial to setup and activate the following jails

  1. Change Defaults per tutorial
  2. nginx-http-auth
  3. nginx-badbots
  4. nginx-nohome
  5. nginx-noproxy

Also make sure the SSH and SSH-DDoS jails are enabled, and consider enabling the recidive filter. I also recommend adding a jail for WordPress via the WP Fail2ban plugin for wordpress, which can be easily installed and activated by following their instructions.

Installing and Configuring PHP 7.2:
Since we are using Ubuntu 18.04, PHP 7.2 is the default for PHP so simply run in terminal

sudo apt install -y zip unzip php-fpm php-mysql php-xml php-gd php-mbstring php-zip php-curl 

Just an FYI that this also installs the MySQL, XML, Curl and GD packages so that WordPress can interact with the database, support for XMLRPC (required for Jetpack), and also automatically cropping and resize images. It also installs zip/unzip because I use zip and unzip in some of my own backup plugins and tools.

I also like to tweak the php.ini settins to allow for more memory and larger file sizes. So let’s open /etc/php/7.2/fpm/php.ini.

sudo nano /etc/php/7.2/fpm/php.ini

You can make this faster by using the search function with CTRL + W and then typing what you’re looking for. I usually recommend increasing the post_max_size from the default 8MB, upload_max_filesize from the default 2MB, and memory_limit from it’s default. I generally set all of mine to 128MB and 256MB respectively

Now let’s restart PHP

sudo service php7.2-fpm restart

Now we need to tell Nginx to use PHP7.2-fpm, so let’s open up our configuration file for our default site.

sudo nano /etc/nginx/sites-available/default

We need to edit the file so that it looks like below, but change and to your TLD that you are using with your server.

server {
  listen 80 default_server;
  listen [::]:80 default_server;
  root /var/www/html;
  index index.php index.html;
  location / {
    try_files $uri $uri/ =404;
  location ~ \.php$ {
    include snippets/fastcgi-php.conf;
    fastcgi_pass unix:/run/php/php7.2-fpm.sock;
  location ~ /\.ht {
    deny all;

Save and exit, then restart Nginx to apply changes

 sudo service nginx restart

Now it’s time to test it all out and make sure this is all working properly. So let’s make a sample PHP file in your /var/www/html folder called index.php

echo "<?php phpinfo();" | sudo tee /var/www/html/index.php > /dev/null

Now open up your web browser and go to http://SERVER.IP.ADDRESS.HERE (e.g., and you should see something like this.

Awesome sauce, we’re starting to see it finally coming together! You officially have made a Linux, Nginx, MariaDB, and PHP stack aka a LEMP stack. Honestly at this point you can serve up just about any LEMP needs you have for any software such as NextCloud or more. Let’s move on, the goal line is within sight!

Encrypt! Encrypt! Encrypt! Let’s Encrypt, with TLS/SSL Certificates from

This is pretty straight forward but I recommend reading Digital Ocean’s tutorial on setting up and securing nginx, to fully grasp what we are doing here. So let’s install letsencrypt. Before you used to have to add a PPA, update, and install certbot, but it’s in the main Ubuntu repo these days so one command to install letsencrypt, and another to install the certificates to the domains defined in the /etc/nginx/sites-available config file as we have done earlier.

 sudo apt install -y letsencrypt
sudo certbot --nginx

Now just follow the instructions, and provided you entered your domains correctly into the Nginx config file, certbot should find and install certificates for all of the domains. Make sure to pick a reliable email for alerts from letsencrypt.

If for some reason certbot can’t find them or you want an SSL for another domain that is pointed at your server you can generate a certificate by using “-d domain.tld” for all the domains you want like so, and bare in mind www.example.tld and example.tld are considered two different domains, so you need to include both in the certificate you generate along with any other subdomains.

 sudo certbot -d example.tld -d www.example.tld

Now we need to edit the Nginx snippet created by certbot.

 sudo nano /etc/letsencrypt/options-ssl-nginx.conf

Edit it so it looks like below, althought the top few lines are created by certbot, so add the ones below to enhance our security profile.

# automatically added by Certbot
ssl_session_cache shared:le_nginx_SSL:1m;
ssl_session_timeout 1440m;
ssl_protocols TLSv1 TLSv1.1 TLSv1.2;
ssl_prefer_server_ciphers on;
ssl_ecdh_curve secp384r1;
ssl_session_tickets off;
ssl_stapling on;
ssl_stapling_verify on;
resolver valid=300s;
resolver_timeout 5s;
add_header Strict-Transport-Security "max-age=63072000; includeSubdomains; preload";
add_header X-Frame-Options SAMEORIGIN;
add_header X-Content-Type-Options nosniff;

Now save and exit.

It’s extremely important to renew your Letsencrypt certificates every couple months at least as they expire every 90 days. So we need to setup a cron job to check for renewals often, and renew the certificates automatically. So lets edit the crontab as root

sudo crontab -e

Add the following lines so we can have it check and autorenew certificates every Monday.

30 2 * * 1 /usr/bin/certbot renew >> /var/log/le-renew.log
35 2 * * 1 /bin/systemctl reload nginx

Now lets save that and run certbot in a dry run to see if renewals will work.

sudo certbot renew --dry-run

Now it’s time to install WordPress
Personally I like to install wp-cli and then finish it up in the WebUI. I love WP CLI as it is a command line interface to administrate wordpress. So if a worst case happens and you say lock yourself out and can’t reset the password, want to install or deactivate a plugin that isn’t allowing wordpress to work, or more it can do it. It’s extremely powerful and handy to have on a system regardless. So let’s install that, then have it download the WordPress files.

curl -O
chmod +x wp-cli.phar
sudo mv wp-cli.phar /usr/local/bin/wp
cd /var/www/html
sudo wp core download

Now go to your domain, and you can run the WordPress quick install, it’s straightforward just enter all the information that it asks for, once that is done hop back into terminal and let’s use WP CLI to install some plugins easily that I recommend with this setup to integrate with the caching on the OS, Fail2Ban, and more. If not planning to use mailgun I recommend gmail-smtp if you use gmail. But do not install both Mailgun and Gmail-Smtp, pick one. I also added Cloudflare because I use that, and it’s a free CDN as well as proxy to help avoid DDoS attacks. They have a free plan that is great.I also added WP-Sweep a great database cleaner tool, and Updraft plus, one of the best WordPress backup software. Plus iThemes Security which I really like for it’s many free security features.

sudo wp plugin delete hello --allow-root
sudo wp plugin install nginx-helper --allow-root
sudo wp plugin activate nginx-helper --allow-root
sudo wp plugin install mailgun --allow-root
sudo wp plugin activate mailgun --allow-root
sudo wp plugin install jetpack --allow-root
sudo wp plugin activate jetpack --allow-root
sudo wp plugin install gmail-smtp --allow-root
sudo wp plugin activate gmail-smtp --allow-root
sudo wp plugin install cloudflare --allow-root
sudo wp plugin activate cloudflare --allow-root
sudo wp plugin install wp-sweep --allow-root
sudo wp plugin activate wp-sweep --allow-root
sudo wp plugin install updraftplus --allow-root
sudo wp plugin activate updraftplus --allow-root
sudo wp plugin install wp-better-security --allow-root
sudo wp plugin activate wp-better-security --allow-root

Mailgun Setup
You’ll need to setup an account. I recommend despite their recommendation, making your domain the same as your regular domain, do not subdomain it. The reason why is you can up a forwarding rule so if say someone emails you at [email protected] it could look professional and forward to a gmail account per se. After you’ve set up your domain at Mailgun, go to Settings > Mailgun from the WP dashboard, copy and paste in your Mailgun domain name and API key, and then click “Save Changes” to get it set up. Click “Test Configuration” to make sure it is working. You may also want to use the Check Email plugin just to make sure that emails are being sent correctly.

GMail SMTP Setup
If you setup the GMail SMTP servers in your DNS according to this guide, you’ll want to have installed the GMail SMTP plugin for WP. The setup for this plugin is somewhat involved. I strongly urge you to follow the instructions on their documentation site.

Time to Optimize and Secure the WordPress

Here are some tips for securing and optimizing your wordpress. Simply replace the content of /etc/nginx/sites-available/default with the following and make sure any reference of “” reflects your actual domain and tld.

fastcgi_cache_path /var/run/nginx-cache levels=1:2 keys_zone=WORDPRESS:100m inactive=60m;
fastcgi_cache_key "$scheme$request_method$host$request_uri";
fastcgi_cache_use_stale error timeout invalid_header http_500;
fastcgi_ignore_headers Cache-Control Expires Set-Cookie;
server {
  listen 80 default_server;
  listen [::]:80 default_server;
  listen 443 ssl http2 default_server;
  listen [::]:443 ssl http2 default_server;
  ssl_certificate /etc/letsencrypt/live/; # managed by Certbot
  ssl_certificate_key /etc/letsencrypt/live/; # managed by Certbot
  include /etc/letsencrypt/options-ssl-nginx.conf; # managed by Certbot
  ssl_dhparam /etc/letsencrypt/ssl-dhparams.pem; # managed by Certbot
  # force redirect to HTTPS from HTTP
  if ($scheme != "https") {
    return 301 https://$host$request_uri;
  client_max_body_size 256M;
  root /var/www/html;
  index index.php index.html;
  set $skip_cache 0;
  if ($request_method = POST) {
    set $skip_cache 1;
  if ($query_string != "") {
    set $skip_cache 1;
  if ($request_uri ~* "/wp-admin/|/xmlrpc.php|wp-.*.php|/feed/|index.php|sitemap(_index)?.xml") {
    set $skip_cache 1;
  if ($http_cookie ~* "comment_author|wordpress_[a-f0-9]+|wp-postpass|wordpress_no_cache|wordpress_logged_in") {
    set $skip_cache 1;
  location ~ /purge(/.*) {
    fastcgi_cache_purge WORDPRESS "$scheme$request_method$host$1";
  location / {
    try_files $uri $uri/ /index.php?$args;
    limit_req zone=one burst=50;
  # Turn off directory indexing
  autoindex off;
  # Deny access to htaccess and other hidden files
  location ~ /\. {
    deny  all;
  # Deny access to wp-config.php file
  location = /wp-config.php {
    deny all;
  # Deny access to revealing or potentially dangerous files in the /wp-content/ directory (including sub-folders)
  location ~* ^/wp-content/.*\.(txt|md|exe|sh|bak|inc|pot|po|mo|log|sql)$ {
    deny all;
  # Stop php access except to needed files in wp-includes
  location ~* ^/wp-includes/.*(?<!(js/tinymce/wp-tinymce))\.php$ {
    internal; #internal allows ms-files.php rewrite in multisite to work
  # Specifically locks down upload directories in case full wp-content rule below is skipped
  location ~* /(?:uploads|files)/.*\.php$ {
    deny all;
  # Deny direct access to .php files in the /wp-content/ directory (including sub-folders).
  # Note this can break some poorly coded plugins/themes, replace the plugin or remove this block if it causes trouble
  location ~* ^/wp-content/.*\.php$ {
    deny all;
  location = /favicon.ico {
    log_not_found off;
    access_log off;
  location = /robots.txt {
    access_log off;
    log_not_found off;
  location ~ \.php$ {
    include snippets/fastcgi-php.conf;
    fastcgi_pass unix:/run/php/php7.2-fpm.sock;
    fastcgi_cache_bypass $skip_cache;
    fastcgi_no_cache $skip_cache;
    fastcgi_cache WORDPRESS;
    fastcgi_cache_valid 60m;
    include fastcgi_params;
  ## Block file injections
    set $block_file_injections 0;
    if ($query_string ~ "[a-zA-Z0-9_]=http://") {
        set $block_file_injections 1;
    if ($query_string ~ "[a-zA-Z0-9_]=(\.\.//?)+") {
        set $block_file_injections 1;
    if ($query_string ~ "[a-zA-Z0-9_]=/([a-z0-9_.]//?)+") {
        set $block_file_injections 1;
    if ($block_file_injections = 1) {
        return 403;
  ## Block SQL injections
    set $block_sql_injections 0;
    if ($query_string ~ "union.*select.*\(") {
        set $block_sql_injections 1;
    if ($query_string ~ "union.*all.*select.*") {
        set $block_sql_injections 1;
    if ($query_string ~ "concat.*\(") {
        set $block_sql_injections 1;
    if ($block_sql_injections = 1) {
        return 403;
  ## Block common exploits
    set $block_common_exploits 0;
    if ($query_string ~ "(<|%3C).*script.*(>|%3E)") {
        set $block_common_exploits 1;
    if ($query_string ~ "GLOBALS(=|\[|\%[0-9A-Z]{0,2})") {
        set $block_common_exploits 1;
    if ($query_string ~ "_REQUEST(=|\[|\%[0-9A-Z]{0,2})") {
        set $block_common_exploits 1;
    if ($query_string ~ "proc/self/environ") {
        set $block_common_exploits 1;
    if ($query_string ~ "mosConfig_[a-zA-Z_]{1,21}(=|\%3D)") {
        set $block_common_exploits 1;
    if ($query_string ~ "base64_(en|de)code\(.*\)") {
        set $block_common_exploits 1;
    if ($block_common_exploits = 1) {
        return 403;

Then while we are at it let’s make sure /etc/nginx/nginx.conf has an additional parameter we set. So open up it up with nano

sudo nano /etc/nginx/nginx.conf

Then look to see if the following block is there under the http section, and make sure it refers to zone one.

http {
    limit_req_zone  $binary_remote_addr  zone=one:10m   rate=2r/s; 

This config file will take advantage of the advanced caching capabilities of our custom version of NGINX. It will also prevent visitors from accessing files that they shouldn’t be. This also adds some configurations to block SQL and file injection attacks, as well as blocking common exploits. Plus we also added some rate limiting so it can help prevent a Denial of Service attack. The combined effect will be to make your site faster and more secure.

Admin’s Have to get Alerts. Set those Admin Emails Up!

Sometimes things happen and you need to know when they happen. So we need to setup email alerts, and while there are a number of ways to do this, this is the best way I recommend to less advanced Linux users. The two ways here will either route through Mailgun, or Gmail depending on what you did earlier will determine what you will do right now. It is based on this tutorial from the EasyEngine folks. First, install the necessary packages. When prompted about your server type, select “Internet Site”, and for your FQDN, the default should be acceptable. Then open the config file for editing:

sudo apt install -y postfix mailutils libsasl2-2 ca-certificates libsasl2-modules
sudo nano /etc/postfix/

We’ll need to edit the “mydestination” property and add a few more, but we can leave the rest as their defaults.

mydestination = localhost.$myhostname, localhost
relayhost = []:587
smtp_sasl_auth_enable = yes
smtp_sasl_password_maps = hash:/etc/postfix/sasl_passwd
smtp_sasl_security_options = noanonymous
smtp_tls_CAfile = /etc/postfix/cacert.pem
smtp_use_tls = yes

If you’re using Gmail as your SMTP server, edit it slightly to look like the following

mydestination = localhost.$myhostname, localhost
relayhost = []:465
smtp_sasl_auth_enable = yes
smtp_sasl_password_maps = hash:/etc/postfix/sasl_passwd
smtp_sasl_security_options = noanonymous
smtp_tls_CAfile = /etc/postfix/cacert.pem
smtp_use_tls = yes
smtp_tls_wrappermode = yes
smtp_tls_security_level = encrypt

Now save that file and it’s time to make a file to store our SMTP credentials

 sudo nano /etc/postfix/sasl_passwd

Now add one of the following single lines, only use one of them, and only pick the one you need for Mailgun or Gmail. Where “PASSWORD” is, of course put your actual password”

[]:587 [email protected]:PASSWORD
[]:465 [email protected].com:PASSWORD

You’ll have to get the password for the postmaster account from your Mailgun dashboard. The password for the GMail example should be the password for the email address used. Next we need to lock down this file and tell postfix to use it by running the following:

$ sudo chmod 400 /etc/postfix/sasl_passwd
sudo postmap /etc/postfix/sasl_passwd
cat /etc/ssl/certs/thawte_Primary_Root_CA.pem | sudo tee -a /etc/postfix/cacert.pem

Now it’s testing time

  sudo /etc/init.d/postfix reload
echo "Test mail from postfix" | mail -s "Test Postfix" [email protected]

If everything went perfect, you’ll receive an email from the server at the address in the last line. Also you can check the mailgun logs to see if it routed through their servers.

FINISH HIM! Auto Updating the server
So we need to make sure our server it automatically applying security updates for obvious reasons. So now we need to enable auto updates for apt.

sudo nano /etc/apt/apt.conf.d/50unattended-upgrades

After editing the file it should look like this

Unattended-Upgrade::Allowed-Origins {
Unattended-Upgrade::Mail "[email protected]";
//Unattended-Upgrade::MailOnlyOnError "true";
Unattended-Upgrade::Remove-Unused-Dependencies "true";
Unattended-Upgrade::Automatic-Reboot "true";
Unattended-Upgrade::Automatic-Reboot-Time "02:00";

This will tell the server to automatically apply security and regular updates, email the admin when updates are done, automatically remove unused dependencies, automatically reboot if necessary, and reboot at 2AM if necessary. But now we need to edit the 10periodic to enable some options.

sudo nano /etc/apt/apt.conf.d/10periodic

Once done it should look like this

APT::Periodic::Update-Package-Lists "1";
APT::Periodic::Download-Upgradeable-Packages "1";
APT::Periodic::AutocleanInterval "7";
APT::Periodic::Unattended-Upgrade "1";

What that tells it to do is run “apt update” to pull new packages, download packages that are available for update, automatically clean package installers weekly, and enable the unattended upgrade we configured prior.

Finally lets do one last update and clean the server up for it’s maiden voyage.

sudo apt-get update
sudo apt-get upgrade
sudo apt-get autoclean
sudo reboot

Another thing that maybe useful is adding a swapfile, I prefer to instead go to a larger server with more RAM as a Swap file isn’t as ideal as more RAM, but better than nothing if you absolutely need it. Digital Ocean has a great tutorial here.

This was a bit of a longer tutorial, and there is a whole lot more you can do from additional wordpress plugins, to a CDN like Cloudflare which can really speed up the site, additional security from Port Scan Attack Detector (PSAD), additional blocklists, and more. I hope to cover an addition to this tutorial in the future to detail how I got multiple website on the same box using a slightly modified version of this stack

Moving to Xubuntu 16.04 for my personal items

Well my Home Theater PC is just about finished… I just installed samba, enabled openSSH via my SSH key, along with my other goodies. Just doing the final tweaks to smooth out video playback via xorg.conf hopefully all the settings will carry over perfectly as this the same exact GPU and such.
But my Nvidia driver is 361 instead of 358 which I was using when on 14.04.
I am just hoping I don’t have to recompile a kernel just for the “Intel Core 2 or Newer” CPU’s along with modifying the system timer from 250Hz to at least 300Hz to have a number more evenly divisible into 30FPS to equal US NTSC video vs European PAL at 25FPS, if that is the case I will bump it to a 1000Hz timer for faster responsiveness as it’s a desktop, as well as being divisible into 30 and 25 in case I take it over seas. If I have to recompile for the Home Theater I am merely going to recompile from the Ubuntu sources rather than directly from My laptop will be getting a custom kernel from though as it doesn’t have the proprietary GPU for smoother video playback

I opted for fresh reinstalls for my laptop and HTPC as I ran into issues with the upgrade and then merely reinstalled all my applications and such through my scripts, as I always keep my /home on a separate partition just in case I have to reinstall things.

I am working on upgrading my servers from 14.04LTS to 16.04LTS as well by Q2 of 2017 after lots of testing. Some servers will probably be rebuilt from scratch. However I am in no rush as we all have until 2019 for the end of life of Ubuntu 14.04 LTS to end.

Nginx Filters for Fail2Ban

When making my other servers I was double checking fail2ban configurations and noticed there is no fail2ban settings for nginx seeing as the webmail runs on it. Not sure if it’s an issue, or anything but I was hoping some other could tell me if I am on the right track, or if it’s not even necessary.

I did this for my email server which runs nginx as the web server.

In the /etc/fail2ban/jail.local

enabled  = true
filter   = nginx-http-auth
port     = http,https
logpath  = /var/log/nginx/error.log
enabled  = true
port     = http,https
filter   = nginx-badbots
logpath  = /var/log/nginx/access.log
maxretry = 2

thencd /etc/fail2ban/filter.d
sudo nano nginx-http-auth.conf

make sure it’s like below


failregex = ^ \[error\] \d+#\d+: \*\d+ user "\S+":? (password mismatch|was not found in ".*"), client: <HOST>, server: \S+, request: "\S+ \S+ HTTP/\d+\.\d+", host: "\S+"\s*$
            ^ \[error\] \d+#\d+: \*\d+ no user/password was provided for basic authentication, client: <HOST>, server: \S+, request: "\S+ \S+ HTTP/\d+\.\d+", host: "\S+"\s*$

ignoreregex =

copy badbots config from apache
sudo cp apache-badbots.conf nginx-badbots.conf

Free(dom) Software: Why Your PC should have Liberty



“Freedom” is the very word at the core of libertarianism: the ability for us to be able to do anything we believe in freely, so long as it does not infringe on the freedoms of another individual. When you think of freedom, what comes to your mind? Speech, religion, assembly, press, property, perhaps the right to bear arms, perhaps the ability for us to do whatever we please to our own bodies, and we can go on from there. So why do many not consider freedom to mean the ability to own our software in our own computers like we view our bodies or our vehicles? This is where Free, Libre, and Open Source Software (FLOSS aka Free or Libre) comes in. What exactly does Free and Open Source mean? Free and Open source software is software whose source code is available for modification or enhancement by anyone. But let us always remember, Free software came before Open Source.

“Source code” is the part of software that most computer users don’t ever see; it’s the code which computer programmers can manipulate to change how a piece of software—a “program” or “application”—works. Programmers who have access to a computer program’s source code can improve that program by adding features to it, or fixing parts that don’t always work correctly.


open source logo


The problems with Windows and Apple are that you can’t really trust them all that much in terms of privacy. They both have backdoors for the government. Now, while the government may have good intentions according to some, they are severely flawed. The problems with a backdoor in a system that I can’t close as a user, means that malicious hackers have another exploit to get into your system with and potentially monitor or steal information from you. Allowing software to be Free means that it is transparent and vetted by thousands of people all around the world, who are constantly working on the software.

Take for example Mozilla Firefox, formerly Netscape Navigator. It’s not maintained by some giant company like Google, Apple, or Microsoft that make their browsers and have parts that are completely closed. Firefox, being Free, is maintained by a large community who make one of the freest browsers, with a non-profit foundation maintaining all the thousands of additions to code that go on every day. We can ensure there are no intentional backdoors in the code, while making it as customizable as we want. In fact the GNU/Linux (often just called “Linux”) back door attempt of 2003 some suspected of being done by the NSA, is proof of that very fact that Free software is more secure.

“Free software” means software that respects users’ freedom and community. Roughly, it means that the users have the freedom to run, copy, distribute, study, change and improve the software. Thus, “free software” is a matter of liberty, not price. To understand the concept, you should think of “free” as in “free speech,” not as in “free beer”. We sometimes call it “libre software” to show we do not mean it is gratis.


Richard Stallman, founder of the Free Software Movement, said Windows and OS X are malware, even stated the observation Amazon’s Kindle has an Orwellian back door, and has said that only an idiot would trust the Internet of Things.

“Malware is the name for a program designed to mistreat its users,” Stallman wrote in The Guardian.

What kinds of programs constitute malware? Operating systems, first of all. Windows snoops on users, shackles users and, on mobiles, censors apps; it also has a universal backdoor that allows Microsoft to remotely impose software changes. Microsoft sabotages Windows users by showing security holes to the NSA before fixing them.

Apple systems are malware too: MacOS [OS X] snoops and shackles; iOS snoops, shackles, censors apps and has a backdoor. Even Android contains malware in a non-free component: a backdoor for remote forcible installation or de-installation of any app.

quote-to-have-the-choice-between-proprietary-software-packages-is-being-able-to-choose-your-master-richard-stallman-268889-768x361 (1)
Richard Stallman: Founder of the Free Software Movement. President of the Free Software Foundation

In fact the entire Free, Libre, and Open Software Software Movements are already unsung heroes of our most precious example of our freedom, the internet. People don’t recall, but many years ago the internet as a whole was very close to becoming proprietary in the software market. You had to buy a lot of closed source software to even connect to the internet. Once upon a time, you actually used to pay for an Internet browser. Microsoft was very close to choking off freedom on the internet when it had, at its peak, 95% of the market share on internet browsers. The internet stagnated for many years, and we were stuck with Internet Explorer, one of the worst browsers ever because it was so full of holes, for a very long time. Microsoft could have also done the same with Windows server,when in fact the majority of web servers and critical components that run the web today are run on GNU/Linux or FreeBSD which are both Free and Open Sourced. In fact GNU/Linux and FreeBSD were the first to have the components to even be able to connect to the internet. For years people were forced to pay for software to create documents, we had Open Office (now deprecated), and LibreOffice which has replaced Open Office. In fact I wrote this rough draft on my laptop running GNU/Linux using LibreOffice. The Being Libertarian site itself is run on GNU/Linux and WordPress, both are Free and Open Sourced software. In fact Bitcoin wouldn’t even exist were it not for FLOSS. I have actually been a member of the FOSS and FLOSS movements for 10 years now, starting when I was 14 years old. GNU/Linux and the FLOSS communities were the first to introduce many of the features now found common on Windows and Mac OS X, including, but not limited to, remote desktop, virtual desktops, and a TCP/IP stack so they can use the internet. In fact a lot of Mac OS X relies on Open Source software, but it’s not as Free as it actually should be, because there is no way I can have access to the complete source code and compile my own Mac OS X, like I can do with GNU/Linux or FreeBSD.

Use-of-Open-Source-Software-Is-Now-Mandatory-In-Indian-Government-Offices-477052-2-768x390 (1)

Now I know a lot of libertarians and conservatives would rather place their trust in businesses managing things. But this is where voluntarism come in, because no one is forcing you to contribute to the code. Far more people are using the software without contributing the software, but if you see a problem or feature that can be added you are more than welcome to do so. This is best explained in Eric Raymond’s “Cathedral and the Bazaar”, of why the FLOSS model of a bazaar with the community voluntarily working together is better than the cathedral approach of Apple or Microsoft. It’s also one of the most influential essays ever written in the IT world.

Eric S. Raymond: One of the key people in the Free, Libre, and Open Source Movement, and author of the “Cathedral and the Bazaar”
Eric S. Raymond: One of the key people in the Free, Libre, and Open Source Movement, and author of the “Cathedral and the Bazaar”

Linux overturned much of what I thought I knew. I had been preaching the Unix gospel of small tools, rapid prototyping and evolutionary programming for years. But I also believed there was a certain critical complexity above which a more centralized, a priori approach was required. I believed that the most important software (operating systems and really large tools like the Emacs programming editor) needed to be built like cathedrals, carefully crafted by individual wizards or small bands of mages working in splendid isolation, with no beta to be released before its time.

Linus Torvalds’s style of development—release early and often, delegate everything you can, be open to the point of promiscuity—came as a surprise. No quiet, reverent cathedral-building here—rather, the Linux community seemed to resemble a great babbling bazaar of differing agendas and approaches (aptly symbolized by the Linux archive sites, who’d take submissions from anyone) out of which a coherent and stable system could seemingly emerge only by a succession of miracles.

*I highly recommend reading the entire essay (~36 Pages on paper)



I have mentioned LibreOffice and Mozilla Firefox as FLOSS software, but there are thousands of distributions of GNU/Linux to use, and thousands of Free, Libre, and Open Source Software out there. It’s all about freedom of choice. If you don’t like one part of the system or program you can change it. For example I used to be a Gentoo GNU/Linux user where I can completely build my own system the way I wanted, then I became a bit lazy and switched to Ubuntu a few years ago as I didn’t want to have to spend so much time on each installation. I, of course, still didn’t like Ubuntu on a few parts, so I modified a lot of the operating system, until eventually Ubuntu decided to add the Unity User Interface, which I hated with a passion. So I switched to Xubuntu as a day to day work OS, withKali GNU/Linux on the same laptop for all my tools to penetrate network security as part of my job. One GNU/Linux in particular a lot of people who love their privacy like is Tails, which leaves no trace of you on the host computer, and encrypts as well as anonymizes all of your data. The Condor intraoral scanner, which I previously covered inMarch of 2015, is running Manjaro GNU/Linux. I also have been looking into Trisquel GNU/Linux which is a completely Free version of GNU/Linux, which Richard Stallman himself uses. There are parts in a quite a few distributions of GNU/Linux that use “non-free” software meaning its use, redistribution or modification is prohibited, or requires you to ask for permission, or is restricted so much that you effectively can’t do it freely. There are a few classes of types of software which the Free Software Foundation does a good job of explaining to the average user.

But back to the main point. The beauty of GNU/Linux and FLOSS in general is freedom, as what works for some, doesn’t work for all. Linus Torvalds, the maker of the GNU/Linux kernel, himself didn’t like Ubuntu, whereas I kind of like the Ubuntu environment.

Linus Torvalds: Founder of the Linux Kernel. In 1991 introduced Linux to the world saying, “I’m doing a (free) operating system (just a hobby, won’t be big and professional like gnu) for 386(486) AT clones.” Despite Linus being the creator of Linux, and also the largest contributor of code to the project, Linus has only contributed 1% of the total code.
Linus Torvalds: Founder of the Linux Kernel.
In 1991 introduced Linux to the world saying, “I’m doing a (free) operating system (just a hobby, won’t be big and professional like gnu) for 386(486) AT clones.” Despite Linus being the creator of Linux, and also the largest contributor of code to the project, Linus has only contributed 1% of the total code.

So at the end of it all, those who call themselves true libertarians, conservatives, or even just a full out privacy rights advocate should be throwing their closed source software away whenever possible, and embrace the true freedom only granted by software to the user which is FOSS or even better, FLOSS. It’s time to really embrace freedom to its fullest extent in all facets of your life. You can even start slowly dipping your toes in the water of the open source movement. I use Mozilla Thunderbird instead of Microsoft Outlook, Mozilla Firefox or even better GNU IceCatinstead of Microsoft Edge or Internet Explorer, Notepad++ (NotepadQQ in GNU/Linux) instead of Notepad,LibreOffice instead of Microsoft Office, GNUCash instead of Quickbooks, ProjectLibre instead of Microsoft Project,7-zip instead of WinZIP and WinRAR, Dia instead of Microsoft Visio, Scribus instead of Microsoft Publisher, and etc. You can find a FOSS or FLOSS solution for almost any task that you need to do today.

download (3)

If any readers are interested in a deep look into the philosophy, culture, and history of the entire Free, Libre, and Open Source Movements, I highly recommend watching the documentary Revolution OS (About 1 1/2 hours) which is freely available on YouTube ( and elsewhere.

Update: After Speaking to Richard Stallman some changes were made

This article was originally written by me for Being Libertarian

Install the latest Mozilla Thunderbird or Firefox in Ubuntu GNU/Linux

So I ran into an issue with my Mozilla Thunderbird today when I was finished setting up my new email, contact, and calendar server with Mail-in-a-box. So I go to add the lightning extension for calendars, and low and behold I find out my Thunderbird (the one that came in the the default Xubuntu repos for 14.04LTS) was out of date and not supported by lightning. The Ubuntu repos had version 38.8, but what version was Mozilla at themselves? 45.1 as of this post. So I quickly installed the latest binary but I tend to be forgetful about updates, so I wanted to tie it into the apt package manager so I found a PPA that works.

First if thunderbird is installed remove it, and maybe backup your .thunderbird folder just in case. But you shouldn’t have to worry about losing any data.

sudo apt-get remove -y thunderbird

Next we need to add a new repository called Ubuntuzilla so edit your sources.list. I used nano for this, but feel free to use whatever you like.

sudo nano /etc/apt/sources.list
add to the end
deb all main

or you can do that all with one command
echo -e "\ndeb all main" | sudo tee -a /etc/apt/sources.list > /dev/null

Then grab the keys and update
sudo apt-key adv --recv-keys --keyserver C1289A29
sudo apt-get update

Install your desired package, with one of the following commands:
sudo apt-get install firefox-mozilla-build
sudo apt-get install thunderbird-mozilla-build
sudo apt-get install seamonkey-mozilla-build


How to make “WHOIS” work with new TLD’s e.g. *.xyz, *.online

So I have been building a lot of servers and generally I like to segment them to different domains but whois by default only will work with *.com, *.info, *.net you know the usual TLD’s you think of. But now there are so many new ones I like to scoop up I still want to test my server settings with whois. Well have no fear on my Xubuntu 14.04LTS I use everyday simply create the file “whois.conf” in the /etc/ folder. So use your favorite text editor and paste this file in to get any new TLD resolved.
Open Nano (or whatever text editor you prefer)
sudo nano /etc/whois.conf

Once inside your text editor paste this list (list is very long so I added a read more section you will need to open to see the entire list)

# WHOIS servers for new TLDs (
# Current as of 2015-09-12

Continue reading “How to make “WHOIS” work with new TLD’s e.g. *.xyz, *.online”

How to Make Super Secure Passwords Easily with One Command

We all know when it comes to security, a secure password is always the most important thing. However remembering a complex password is always the toughest part anywhere. Especially when it comes to being a system administrator, our passwords are usually the most vital of anyone in the company. When it comes to telling people they need complex passwords, what always comes to mind is this xkcd comic about passwords.


As the bottom text suggests we have come to the point where it’s hard for us to remember passwords, but easy for computers to guess. So what’s the solution? Well what I do as a GNU/Linux person is use the command already built in to generate super secure passwords using the sha1sum, sha224sum, sha256sum, sha384sum, and sha512sum commands.

First off pick a random word or phrase. Now remember capitalization, spaces, and such will always effect the sum spit out. let’s start with sha1sum which is the shortest, and using the word “password” as our example throughout this tutorial

echo "password" | sha1sum

So we see using the word “password” it spits out the sha1sum of the word, and we now have a very complex password. Now let’s try it with SHA256

echo "password" | sha256sum

So you see as we increase the strength of the sha256sum, the output sum is longer, and with a longer password comes even more security. Now let’s try SHA512

echo "password" | sha512sum

So now we see the output is incredibly long and complex. This is a great way to make incredibly secure passwords.



Redis Caching with OwnCloud

While setting up an OwnCloud server for my company, I couldn’t really find any good way to cache, and with the Ubuntu repos having an old version of Redis, meant of course it couldn’t be used for best performance and stability. I tried installing it manually from some guides I found, and trying to see OwnCloud’s documentation and was last using an Apcu and Redis (older version) combined so I stumbled upon a guide from which actually resolved my issues of an old Redis, and dramatically sped up my server.

This guide is also scripted for an automated install, you can download the script here.

    $~: sudo php5dismod apcu && sudo apt-get purge php5-apcu -y
    $~: rm /etc/php5/mods-available/apcu-cli.ini
    $~: sudo apt-get purge --auto-remove memcached -y && php5dismod memcached
    $~: sudo apt-get update && sudo apt-get install build-essential -y

Continue reading “Redis Caching with OwnCloud”

Tip for OwnCloud

I was building my OwnCloud file storage on Ubuntu 14.04LTS (upgrading to 16.04.1 LTS this summer), which if you haven’t heard of definitely check out it is the most amazing cloud storage program and you control it yourself. It even offers server side encryption, and tons of options to make it how you want it for you or your company. See it at

But I was coming across an .htaccess issue that kept popping up so I modified Apache so much and it still appeared. So I finally stumbled across my fix. Move the OwnCloud data directory out of the default location. So here are the steps I took

Stop apache2

sudo service apache2 stop

Edit config file in default location

sudo nano /var/www/html/owncloud/config/config.php

Change default location to new location

(pick one, I chose /mnt/owncloud_data but put it anywhere you like)

Move the data folder to new location

sudo mv /var/www/html/owncloud/data /new/data/directory/here

if required change permissions

sudo chown -R www-data:www-data /new/data/directory/here

Restart apache2

sudo service apache2 start

Voila .htaccess issue is GONE!