by myCREcloud | Apr 20, 2026 | Company
The AI boom has quietly become a cloud infrastructure boom. As companies race to roll out AI tools, the biggest line items in tech budgets are no longer flowing into software. They are pouring into chips, servers, power systems, and the massive data centers needed to run large models at scale.
The numbers keep getting bigger
According to figures cited by Reuters, Alphabet, Amazon, Meta, and Microsoft are collectively on pace to spend around $650 billion on AI-related infrastructure in 2026. That is up from roughly $410 billion in 2025, a leap that shows just how quickly the center of gravity in tech spending has shifted toward physical capacity.
Other efforts are adding to the pile. Stargate, a joint project backed by OpenAI, SoftBank, and Oracle, aims to put as much as $500 billion into AI infrastructure across the United States over the coming years. On the venture side, the Stanford AI Index Report pegged global private investment in generative AI at $33.9 billion for 2024, an 18.7 percent jump over the year before.
The bottleneck has moved
Training and running modern AI models is enormously compute intensive. It typically means thousands of GPUs working in parallel across distributed facilities, which makes the networking between chips matter nearly as much as the chips themselves.
That context helps explain Nvidia’s recent decision to commit $2 billion each to photonics firms Lumentum and Coherent. Photonics uses light instead of electrical signals to move data, which can be faster and more power efficient. As AI clusters grow, those internal links increasingly dictate how quickly a system can actually learn and serve predictions.
The lesson for the industry is clear. For most teams building AI, the limiting factor is no longer code. It is whether the underlying hardware, facilities, and power supply can keep up.
What this means for enterprises
Enterprise adoption is a major piece of the story. Businesses are folding AI into customer support, analytics, and internal productivity tools, and very few of them can host that workload on their own servers. They rent capacity from cloud providers instead, often through multi-year commitments worth billions.
For IT leaders, this shifts how cloud decisions get made. GPU availability, data center location, energy supply, and long-term compute costs are all becoming central to vendor evaluations. The next chapter of cloud computing will be shaped less by clever software and more by who has the physical capacity to run it.
Want to learn more about the cloud? Contact Tyler at Tyler.mathis@mycrecloud.com or 619.704.2969!
by myCREcloud | Apr 3, 2026 | Company
Your server is the backbone of your business. It runs your applications, stores your data, and keeps your team productive. When it starts to fail, everything slows down or worse, stops completely.
The good news is that you do not need to be highly technical to perform a basic health check. Here is a straightforward walkthrough of the key areas you should review.
1. Check Your Server License Status
Start with your server’s licensing.
Make sure your operating system and any critical software are properly licensed and up to date. Expired or unsupported licenses can prevent updates, expose you to security risks, and even cause compliance issues.
If you are unsure, you can:
- Check your license key or tag in system settings
- Confirm renewal dates
- Verify support status with your vendor or IT provider
2. Monitor Server Temperature
Heat is one of the biggest threats to server health.
Optimal temperature range:
Between 64°F and 81°F (18°C to 27°C)
Warning signs of temperature issues:
- Server room feels hot or poorly ventilated
- Fans constantly running at high speed
- Unexpected shutdowns or restarts
Suboptimal temperatures, especially prolonged heat, can significantly shorten the lifespan of your hardware.
3. Evaluate Overall Performance
A slow server is often the first noticeable sign of trouble.
A good real world indicator is how your business applications perform, especially Sage.
Ask yourself:
- Is Sage taking longer to open or load data
- Are reports slower than usual
- Are users experiencing lag or freezing
If performance is degrading, it may point to resource constraints, aging hardware, or underlying system issues.
4. Check Storage Capacity
Running out of space can cripple your server.
Look at:
- Total storage capacity
- Percentage of space used
- Growth rate of your data
Best practice:
Keep at least 15 to 20 percent of storage free at all times.
When drives get too full:
- System performance drops
- Backups may fail
- Applications like Sage can become unstable
5. Listen for Unusual Noises
Your server can tell you a lot if you listen.
Be aware of:
- Clicking or grinding sounds from hard drives
- Loud or inconsistent fan noise
- Any new or unusual sounds
These can indicate hardware failure, especially with older spinning drives.
6. Review Your Operating System
Ensure your server operating system is current, supported, and receiving regular security updates.
Verify the following:
- The OS version currently in use
- Whether the OS is still within vendor support lifecycle
- That security patches and updates are actively being released
Risks of running an outdated operating system:
- Increased exposure to cyberattacks
- Compatibility issues with modern applications and services
- Reduced system stability and performance
Important note:
Windows Server 2016 will reach end of extended support on January 12, 2027. After this date, Microsoft will no longer provide security updates, patches, or technical support.
Recommendation:
Begin planning an upgrade or migration before the support deadline to avoid security and compliance risk
7. Consider the Age of Your Server
Age is one of the most important factors.
If your server is more than 5 years old, it should be seriously evaluated for replacement.
Why this matters:
- Hardware performance declines over time
- Failure rates increase significantly
- Warranty coverage is usually expired
- Modern software demands more resources
Your server is not just another piece of equipment. It is your business’s lifeline. Waiting until it fails can lead to costly downtime and lost data.
Schedule a free call with Tyler →
Or request a Sage hosting quote →
by myCREcloud | Apr 2, 2026 | Company
INDUSTRY INSIGHT
The global technology supply chain is under pressure like never before. Here’s what’s driving costs up, and how mycrecloud protects you from the fallout.
If you’ve noticed conversations in the industry about rising infrastructure costs, you’re not imagining things. Across the world, businesses that rely on cloud hosting are grappling with the same reality: the hardware that powers modern computing is more expensive, more scarce, and harder to source than it was just a few years ago.
At myCREcloud, we host Sage solutions for hundreds of businesses across the US. We want to be transparent about what’s happening in the market and why choosing a specialist, committed cloud partner matters more than ever.
- 43% Average server cost increase since 2022
- 2–3× Lead times for enterprise hardware
- Bllions | Global AI chip demand surge
The AI boom is consuming the world’s chips
The extraordinary rise of artificial intelligence has fundamentally disrupted the semiconductor market. Hyperscale cloud providers such as Microsoft, Google, Amazon, Meta are purchasing graphics processing units (GPUs) and high-performance CPUs in staggering volumes to power their AI training infrastructure. This has created a cascading effect throughout the entire supply chain.
Even conventional server hardware, the kind that reliably runs business-critical applications like Sage has become harder to source and significantly more expensive. Manufacturers are prioritising AI-adjacent components, leaving the broader enterprise market squeezed.
“A single AI data centre build can consume more server components than an entire country’s enterprise IT refresh cycle and that pressure ripples down to every hosting provider.”
Supply chain fragility hasn’t gone away
The disruptions that began during the pandemic years fundamentally exposed the fragility of global semiconductor supply chains. While some normalisation occurred, the underlying vulnerabilities remain. Much of the world’s advanced chip manufacturing is concentrated in a small number of facilities in East Asia, making geopolitical tensions, particularly around Taiwan — a persistent risk factor that infrastructure pricing reflects.
Energy and raw material costs have also remained elevated. The metals and rare earth elements used in server components — copper, cobalt, lithium — have seen sustained price increases. These costs compound at every stage of the supply chain and eventually land in the price of a rack of servers.
Energy costs: the hidden hardware multiplier
Running modern hardware isn’t just about buying it; it’s about powering it, cooling it, and maintaining it. Energy prices across Europe have remained volatile following the disruptions of 2022, and data centres are energy-intensive by nature. The cost of electricity, cooling infrastructure, and power redundancy systems all factor into what it takes to keep your Sage environment running reliably 24/7.
Responsible hosting providers — those who don’t cut corners on redundancy or uptime guarantees absorb these costs in their infrastructure investment. That investment is what keeps your business running when everything else is under pressure.
What this means for businesses running Sage
For companies hosting Sage 100 Contractor, Sage 300, or Sage Estimating on-premise or with a budget provider, these market pressures represent real risk. Under-resourced hosting environments cut corners on hardware refresh cycles, leaving you running on aging infrastructure with higher failure rates and slower performance. Security patching gets delayed. Redundancy gets reduced.
This is precisely where mycrecloud is different. Rather than passing unpredictability on to you, we:
Plan our hardware procurement years in advance, locking in pricing and availability before market spikes hit
Maintain dedicated Sage-optimized infrastructure, not oversubscribed generic cloud pools
Guarantee predictable, transparent pricing so you can budget with confidence
Provide enterprise-grade redundancy, UK-based data centres, and Sage-accredited support as standard
Specialist hosting: now more valuable than ever
When the market is under pressure, generalist providers reduce quality to protect margins. Specialist providers like myCREcloud do the opposite; we double down on what we know best. Our entire infrastructure exists for one purpose: to give Sage users the fastest, most reliable, most secure cloud experience possible.
The complexity of today’s hardware market is an argument for choosing your hosting partner carefully, not for going it alone or picking whoever is cheapest today. Reliability and expertise, delivered consistently over years, is what protects your business when conditions are difficult.
Ready to move your Sage to the cloud?
Talk to our team about a tailored hosting solution. No jargon, no hidden costs — just reliable Sage hosting from specialists who understand your software.Get in touch with mycrecloud ↗
by myCREcloud | Mar 12, 2026 | Company
Thirty years of lessons, and why they matter right now.
Almost no one believed in Cloud Computing at first. In the late 1990s, the idea that businesses would hand their data and software to a remote server, managed by someone else, seemed absurd. Critics laughed. Analysts hedged. CIOs said it would never fly.
Then it did. And it changed almost everything.
It also went by a dozen names along the way. On Demand. Utility Computing. Application Service Providers. Alternative Delivery Models. Web Services. Software as a Service. The technology found its footing long before it found its identity. Sound familiar?
We are in a similar moment with AI. The hype is enormous. The skepticism is real. The terminology is all over the place. And nobody quite knows how it will reshape the world, only that it will.
Ben Pring has a front row seat to both eras. He wrote the first analyst notes on Cloud Computing in 1997, produced the first market forecast of the Cloud’s growth potential in 2002, and has spent decades advising both buyers and sellers through the noise. In his recent piece, he draws eleven lessons from the Cloud story that he believes will help leaders navigate what is coming with AI.
Here is what stands out.
The Technology That Wins Is Rarely the One That Arrives First
Cloud Computing was not a sudden invention. It built on decades of mainframe thinking, client-server models, and early internet infrastructure. What changed was the moment when the cost, the connectivity, and the cultural readiness all lined up at once.
AI is the same. The algorithms behind today’s large language models are not new. What is new is the compute power, the data availability, and the interface design that made it accessible to everyone. The lesson: do not mistake the moment of mainstream arrival for the moment of invention. There is always more history than the headlines suggest.
Skepticism Is Not the Same as Being Wrong
Early Cloud skeptics were not fools. Their concerns about security, reliability, and vendor lock-in were legitimate. Many of those concerns took years to resolve. Some are still being resolved today.
The same will be true of AI. People who raise concerns about accuracy, bias, intellectual property, and workforce displacement are not technophobes. They are identifying real problems that the industry will need to solve. The difference between a healthy skeptic and someone who gets left behind is whether they stay engaged with the answers as they develop.
The Names Will Keep Changing. The Direction Will Not.
When Salesforce launched in 1999, it did not call itself a Cloud company. When AWS launched in 2006, the term Cloud was still niche. The category found its name years after the category existed.
Today we argue about whether to call things AI, Generative AI, Large Language Models, or Agentic AI. None of it matters much. The underlying shift, machines taking on cognitive work, is the thing to track. Do not let the branding debates distract you from the substance.
The Platform Wars Are Coming. Pick Carefully.
Cloud did not stay fragmented for long. It consolidated around a small number of dominant players: AWS, Azure, Google Cloud. Everyone else either niched down, got acquired, or disappeared.
The AI landscape will likely follow a similar path. Right now it feels wide open. In five years it will probably not be. Businesses making deep integrations today are placing bets whether they realize it or not. Understanding which platforms are likely to persist is one of the most important strategic questions of the next few years.
The Real Value Is in What Gets Built on Top
Cloud infrastructure was not the destination. It was the foundation. The value came from what companies built once they had access to scalable, affordable computing: Netflix, Uber, Airbnb, Spotify, and thousands of others that would have been impossible in a pre-Cloud world.
AI will work the same way. The models themselves are infrastructure. The interesting question is what becomes possible once that infrastructure is widely available and cheap. The companies that figure that out first will define the next decade.
The Lesson Beneath All the Lessons
Pring’s broader point is one worth sitting with. The Cloud era took about 25 years to fully mature. It was messy, non-linear, and full of false dawns and unexpected turns. The companies and leaders who navigated it best were not the ones who predicted every twist. They were the ones who stayed curious, stayed engaged, and updated their thinking as the evidence changed.
That is the actual skill. Not predicting the future. Staying oriented within it.
The AI era will reward the same qualities. Read the full article for all eleven lessons. It is worth your time.
by myCREcloud | Feb 26, 2026 | Company
There’s a phrase that comes up in almost every conversation about cloud adoption usually from someone in IT leadership, arms crossed, coffee going cold: “We know we need to move. We’re just not ready yet.”
It’s understandable. Cloud migrations feel big. They touch infrastructure, workflows, budgets, and people. But here’s the thing: the companies saying “eventually” are quietly falling behind the ones that said “let’s start small and figure it out as we go.” And the gap is growing faster than most leaders realize.
So let’s talk about where cloud adoption actually stands in 2025, why hesitation has gotten more expensive, and what the companies getting it right are doing differently.
The “Wait and See” Window Has Closed
A few years ago, holding off on cloud migration was a reasonable call. The tooling was immature, the security concerns were legitimate, and there were real questions about whether the ROI would pan out. Waiting made sense.
That window is closed.
Cloud infrastructure has matured dramatically. Security frameworks like SOC 2, ISO 27001, and zero-trust architectures have made cloud environments — in many cases — more secure than legacy on-prem setups. The major providers have invested billions into compliance, uptime, and tooling. And the SaaS ecosystem has developed around cloud-native assumptions, meaning that integrations, APIs, and partner tools are all built expecting you to be there.
If your team is still running core operations on legacy infrastructure, you’re not just missing out on efficiency gains. You’re actively working against the grain of how modern software is built.
What’s Actually Holding Teams Back
The reasons for slow adoption have shifted. It’s rarely a technology problem anymore. More often, it’s one of these three things:
Organizational inertia. The systems work (mostly). Changing them means training, disruption, and risk. Nobody wants to own a migration that goes sideways. So it gets kicked to the next quarter, then the next.
The “big bang” misconception. A lot of teams think cloud migration means a massive, all-at-once overhaul. Lift and shift everything. Rebuild the infrastructure. Take a deep breath and flip the switch. This is almost never the right approach, and the mental weight of that imagined project keeps teams stuck.
Unclear ownership. Cloud migration lives in a weird space — it’s a technology project, a finance conversation, and an operations initiative all at once. When it’s everyone’s problem, it’s no one’s priority.
What the Companies Getting It Right Are Doing
The businesses making real progress on cloud adoption share a few things in common, and none of them involve massive upfront commitments or heroic IT projects.
They start with a workload, not a strategy. Rather than trying to define a five-year cloud roadmap, they pick one thing — a reporting tool, a data pipeline, a customer-facing application — and move that first. The learning from that first migration shapes everything that comes after.
They treat cloud costs like a product decision, not a utility bill. Cloud spend is variable, which is new for most finance teams used to predictable infrastructure budgets. The companies that thrive are the ones that actively manage and optimize spend rather than just paying the invoice each month. FinOps as a discipline has grown up fast for exactly this reason.
They upskill continuously, not all at once. Big training initiatives have a way of not sticking. The teams with real cloud fluency built it through hands-on work, small wins, and a culture that treats cloud literacy as an ongoing investment rather than a one-time certification push.
They accept imperfection. Cloud-native is a direction, not a destination. You don’t have to have everything containerized and serverless and perfectly optimized on day one. Progress matters more than purity.
The Real Cost of Waiting
Here’s what doesn’t show up in the budget line for “infrastructure — current year”: the compounding cost of technical debt, the talent you lose to companies with better tooling, the integrations you can’t build because your systems don’t support modern APIs, and the speed you give up every time a new initiative has to work around legacy constraints.
Cloud adoption isn’t just about infrastructure. It’s about what infrastructure enables — faster product iteration, better data, more scalable customer experiences, and teams that spend their time on things that actually matter.
The companies that made the move — even imperfectly, even incrementally — are operating with a structural advantage now. And it compounds.
A Practical Starting Point
If you’re reading this and feeling the gap between where your organization is and where it needs to be, here’s a simple way to start:
Pick one process or system that causes frequent pain. Something that’s slow, brittle, or hard to scale. Ask what it would take to move just that piece to the cloud. Don’t try to solve everything. Just solve that one thing, learn from it, and go from there.
The cloud doesn’t have to be a big leap. It can be a series of small ones — each one making the next a little easier.
That’s how the companies getting it right actually got there.
by myCREcloud | Feb 5, 2026 | Company
The start of a new year is more than a calendar reset—it’s an opportunity to evaluate what’s working, eliminate inefficiencies, and set your construction business up for smarter growth. If Sage 300 Construction and Real Estate (CRE) is at the core of your operations, January is the ideal time to ensure your system is fully optimized to support your financial, project management, and reporting goals for 2026.
Many firms implement Sage 300 CRE with strong intentions but gradually fall into workarounds, outdated workflows, and underutilized features. Over time, these gaps can slow productivity, increase risk, and limit visibility into business performance. A proactive system review now can prevent costly headaches later.
Here’s where to focus your optimization efforts this year.
1. Start with a System Health Check
Before introducing new processes, take a step back and evaluate your current environment.
Ask yourself:
- Are all modules configured to match how your business operates today?
- Are teams relying on spreadsheets outside the system?
- Have reporting needs evolved beyond your current setup?
- Is your infrastructure keeping pace with security and performance expectations?
Even high-performing organizations discover hidden inefficiencies during a formal review. Identifying them early allows you to streamline operations before peak construction season begins.
Pro tip: Look for duplicated data entry, manual approval processes, and delayed reporting—these are often the biggest indicators that your system needs refinement.
2. Standardize Workflows Across Teams
Inconsistent processes create confusion, slow onboarding, and introduce unnecessary risk. Whether it’s job costing, change order approvals, or AP workflows, standardization improves accuracy and accountability.
For 2026, prioritize:
- Documented procedures for accounting and project teams
- Automated approval workflows
- Consistent naming conventions and job structures
- Clear audit trails
When everyone follows the same playbook, leadership gains cleaner data—and cleaner data drives better decisions.
3. Automate Where It Matters Most
Construction firms are under constant pressure to do more with less. Automation is no longer a luxury; it’s a competitive advantage.
Look for opportunities to automate:
- Invoice routing and approvals
- Payroll processes
- Financial reporting
- Compliance documentation
- Data integrations between platforms
Automation reduces human error, accelerates timelines, and frees your team to focus on higher-value work instead of administrative tasks.
If your staff is still chasing paperwork or rekeying data, this is the year to change that.
4. Strengthen Security and Accessibility
Cyber threats continue to rise, and construction companies are increasingly targeted due to the financial data they manage. At the same time, teams expect secure access from anywhere—jobsites, satellite offices, and home.
Modernizing your Sage 300 CRE environment can help you:
- Protect sensitive financial information
- Improve disaster recovery readiness
- Enable secure remote access
- Reduce reliance on aging on-premise servers
A secure, cloud-hosted environment not only mitigates risk but also enhances system performance and reliability.
5. Invest in Training and User Adoption
Even the best-configured system delivers limited value if employees aren’t using it effectively.
As processes evolve, make training a priority:
- Offer refresher sessions for existing staff
- Build structured onboarding for new hires
- Highlight lesser-known features that improve efficiency
- Encourage teams to retire outdated habits
Organizations that treat training as an ongoing strategy—not a one-time event—see stronger adoption and faster ROI.
6. Align Technology with Business Goals
Your technology should support where your company is headed, not where it was three years ago.
Planning to scale? Enter new markets? Increase project volume? Improve cash flow visibility?
Your Sage 300 CRE setup should reflect those ambitions.
Consider partnering with experts who understand construction workflows and can recommend enhancements tailored to your growth strategy. A forward-looking approach ensures your system evolves alongside your business.
Make 2026 Your Most Efficient Year Yet
Optimization isn’t about reinventing your operations overnight—it’s about making intentional improvements that compound over time.
By reviewing your system, standardizing workflows, embracing automation, strengthening security, and investing in your people, you position your organization for a smoother, more profitable year.
The firms that win in today’s construction landscape aren’t just working harder—they’re working smarter with technology that supports them every step of the way.
New year, new processes, stronger foundation. Now is the time to ensure your Sage 300 CRE environment is ready for everything 2026 has in store.