The New Year offers new opportunities and your IT needs to be ready to take advantage. Join our free webinar on Tuesday, November 11 at 1PM and get ahead of your planning.

2025 Honoree Best Places to Work award badge with C&P Business Marketing logo in blue and black
Person in dark clothing working on laptop surrounded by multiple monitors displaying code and data.

System+Signal: The Industrial Revolution of Cybercrime

In the late 18th century, an English textile worker could weave a bolt of cloth in a week. By the mid-19th century, a steam-powered loom could turn out ten times that amount in a day. What had once been the work of artisans was suddenly industrialized. Entire towns changed overnight. Jobs disappeared, new fortunes were made, and an unstoppable tide of production reshaped the world.

Something similar is happening now—not with looms, but with laptops. Cybercrime, once the domain of clever individuals hunched over keyboards, has become a mechanized, industrial-scale operation. The "steam engine" of this revolution is artificial intelligence.

A recent report from Huntress pulled back the curtain on what a modern cybercriminal's operation actually looks like. Forget the hoodie-clad loner in a basement. Instead, imagine an office floor full of cubicles—teams managing stolen credentials, customer support desks handling victims who've been locked out of their systems, and marketing funnels designed to maximize ransom payments. It looks disturbingly like a real business. And now, with AI in the mix, that business is about to scale in ways we're only beginning to grasp.

For years, most small business owners pictured hackers as solitary geniuses. The Huntress team's look inside attacker operations makes clear: today's cybercriminals run more like startups. They track KPIs, hire contractors, and use collaboration tools. They outsource parts of their workflow to other "vendors" in the dark web economy. What AI adds is scale. Where once a phishing campaign required hours of manual effort to write believable emails, today an AI tool can generate hundreds of variants in minutes—in multiple languages, customized to specific industries, and tailored to mimic the tone of a trusted vendor. The industrial loom has arrived.

In New England, the shift is already being felt. Earlier this year, the Littleton Electric Light & Water Department, a small Massachusetts utility, was targeted by a sophisticated cyber-espionage campaign linked to the Volt Typhoon group (American Public Power Association, 2024). Utilities like Littleton's are not high-profile Fortune 500s, but their systems are vital, and the attack forced the organization to overhaul its cybersecurity playbook: segmenting IT and operational networks, increasing password rotation, and deploying vulnerability management tools. It was a wake-up call that even small organizations are now squarely in the crosshairs.

Closer to Boston, a mid-sized financial services firm was hit with a ransomware attack that began with a single phishing email. An employee clicked what looked like a routine message, setting off a chain of compromise. The breach cost the firm more than $300,000 in downtime and recovery, not counting the loss of client trust (Rutter Networking Technologies, 2024). The firm had assumed that "serious" cybercrime only targeted the giants of Wall Street. Instead, AI-fueled automation made them just as attractive a target as any large competitor.

These examples are not isolated. Many of the most advanced cybercriminal groups are not simply rogue actors, but arms of nation-states. China has been linked to espionage groups like Volt Typhoon and APT41, using hacking as an extension of industrial policy to gather intellectual property and disrupt infrastructure. Russia has turned cybercrime into both a revenue stream and a geopolitical tool, with ransomware gangs like Conti and REvil operating with tacit state approval. And North Korea has made cybercrime one of its largest sources of foreign currency, running ransomware campaigns, cryptocurrency thefts, and fraud rings through state-backed teams like Lazarus Group. These aren't hobbyists—they are salaried employees, often working nine-to-five shifts, targeting businesses around the world with industrial-scale precision.

When you combine nation-state resources with AI tools, the scale is staggering. Instead of a criminal gang launching a few hundred phishing emails, entire teams can launch millions of highly personalized attacks in a single campaign. Deepfake voices can target executives in dozens of languages simultaneously. Malware can be updated and redeployed in hours, not weeks. For small businesses in Massachusetts, it means you are not just defending against a random "hacker." You are up against global organizations backed by countries with deep pockets and strategic motives.

National data reinforces just how widespread this has become. A survey by Nationwide found that one in four small business owners has already been targeted by an AI-driven scam—whether through voice, video, or email impersonation (Nationwide, 2024). The reason is simple: AI has lowered the cost of attacks and raised their believability. Criminals—and the nation-states backing them—can now afford to cast a wider net, and SMBs are in that net.

The comparison to the industrial revolution is not just metaphorical. Just as machines made textiles cheaper, faster, and available to the masses, AI-driven cybercrime makes attacks cheaper to launch, faster to scale, and widely accessible to criminals with little technical skill. For small and mid-sized businesses, this changes the economics of risk. It is no longer a matter of if you will be targeted, but how often—and whether your defenses scale as effectively as the attacks against you.

The natural question is whether defenders can use AI in return. To some degree, yes. Modern security platforms use machine learning to detect unusual behavior, like a login from Boston at 9 a.m. followed by another from Moscow at 9:05. These tools give small businesses a fighting chance against the scale of automated attacks. But business owners should be skeptical of sweeping promises. AI isn't a magic shield. What matters is the system around it: patching, monitoring, backups, and above all, employees who are trained to pause when something feels off.

If there's one advantage businesses still have, it's people. AI may be able to mimic language, generate invoices, or replicate a voice, but it cannot replace the intuition of a well-trained employee who double-checks an account number or insists on written confirmation before wiring funds. Policies and culture matter as much as software. In many of the breaches we've seen, a single human moment of hesitation could have prevented catastrophe.

For professional services firms, medical practices, manufacturers, and local utilities alike, the lesson is the same: the machine is already running. Just as 19th-century looms churned out fabric day and night, the AI-driven cybercrime engine hums along without pause. For criminals, it's not personal. It's not about you. It's about efficiency. You're not a victim in their eyes—you're throughput.

Business owners have a choice, though. The industrial revolution reshaped economies, but it also reshaped safety standards, business practices, and the way work was organized. The AI revolution in cybercrime demands the same. Not a scramble for the latest shiny tool, but a thoughtful approach that blends human awareness with technical safeguards. Resilience comes not from a single product, but from layered defenses, clear policies, and partners who understand both the technology and the stakes.

When the first steam-powered looms appeared, many dismissed them as curiosities. Within a generation, they had transformed society. AI-driven cybercrime is on a similar trajectory. Business owners who shrug it off as hype today may find themselves woven into someone else's profit model tomorrow. The loom is running. The question is whether your business is prepared for the fabric it's weaving.