Skip to main content
59
results
Andrew Fletcher
Building a resilient web server   Early in the year, we encountered a number of challenges, one of the most disruptive was a sustained wave of automated attacks hammering several of the web servers we manage. The volume of traffic overwhelmed server resources, slowing websites to a crawl and frequently triggering 504 Gateway Timeout errors. With uptime and user experience at risk, we needed a robust solution to detect and block malicious traffic in real time, before it could do further...
Andrew Fletcher
Across every industry, the operational risks of cyber threats are escalating....
Andrew Fletcher
Securing sensitive content is essential for protecting data integrity and user...
Andrew Fletcher
When working on Drupal projects, especially in a collaborative environment, it’s...
Andrew Fletcher
If you're working with Drupal on an Ubuntu 24.04 server, it's common to use the...
Andrew Fletcher
Maintaining a Drupal site is usually smooth sailing once the environment is properly set up. However, even in a stable environment, updates to modules can sometimes reveal underlying configuration issues that weren't apparent before. This was the case when I updated a contrib module on a Drupal 10.3 site, which suddenly started throwing errors related to the mbstring PHP extension.   The problem The problem first surfaced after updating the search_api_attachments contrib module. The Drupal...
Andrew Fletcher
If you’ve ever encountered strange errors while running a bash script, you’re...
Andrew Fletcher
Have you ever wanted to use Terminal with a more fluid style and increase your...
Andrew Fletcher
When it comes to text editing, Visual Studio Code (VS Code) is one of the most...
Andrew Fletcher
On a server directory with over 100,000 PDF files, I needed to verify the...
Andrew Fletcher
Managing large directories, especially those filled with a diverse mix of media such as documents and images, can quickly become cumbersome. In my latest project, dealing with a directory that totals a whopping 87GB, the need for efficient file compression and selective archiving became apparent. How can we streamline this process using the zip command, particularly when we want to exclude non-essential files and directories?   The routine of zipping files in Linux In Linux, zipping files...
Andrew Fletcher
Recently I came across this piece of gold when dealing with databases,...
Andrew Fletcher
In the realm of system administration and monitoring, understanding memory usage...
Andrew Fletcher
The red flag was when I saw the server disk space is showing a site is taking up...
Andrew Fletcher
To determine the size of a directory using the terminal, you can use the du...