📞 The Call - Case Opened
Date: January 30, 2024, 11:30 AM Client: MediaVerse Studios (Media & Entertainment Company) Contact: Jordan Lee, Director of Cloud Operations
“Detective, our S3 storage costs are out of control. We’re a media company, so we expect high storage needs, but $12,350 per month seems excessive even for us. Our content teams swear they need everything, but I suspect there’s a lot of forgotten data. Can you investigate without disrupting our production workflows?”
A classic case of digital hoarding—with potential for substantial savings.
🔍 Initial Investigation - Storage Census
The Storage Landscape
MediaVerse’s S3 footprint was considerable:
- 156 buckets across 2 regions
- 97.4 TB total storage
- $12,350/month in S3 costs
- 3 storage classes in use (Standard, IA, One-Zone IA)
Most concerning: No Glacier storage despite being a media company with archival needs.
First Signs of Waste
Our S3 Storage Analyzer flagged concerning patterns:
SUSPICIOUS STORAGE PATTERNS:
- 23.2 TB in Standard tier with no access for 2+ years
- 45 buckets with "temp", "old", or "archive" in names
- No lifecycle policies on 143/156 buckets
- 17.6 TB of non-compressed video in Standard storage
🕵️ Deep Investigation - Access Patterns & Metadata
The Telling Timestamps
The evidence was overwhelming:
Evidence File #1: Last Access Analysis
Access Patterns for Flagged Data (23.2 TB):
- Last GET request: >730 days ago
- Last PUT request: >730 days ago
- Last LIST request: >90 days ago
- Pattern: Complete inactivity
Evidence File #2: Content Age Analysis
Content Creation Timeline:
- 8.4 TB: 2-3 years old
- 11.7 TB: 3-5 years old
- 3.1 TB: >5 years old
Evidence File #3: Financial Impact
Current Storage Costs (23.2 TB):
- S3 Standard: $0.023 per GB/month
- Monthly cost: $546.51 * 15 = $8,197.65
Projected Glacier Costs (23.2 TB):
- Glacier Deep Archive: $0.00099 per GB/month
- Monthly cost: $23.61 * 15 = $354.15
Potential monthly savings: $7,843.50
🔎 The Evidence Trail - What Was This Data?
Content Inventory Discovery
A metadata analysis revealed the nature of the forgotten content:
Completed Projects (14.7 TB):
- Marketing campaigns from 2018-2020
- Completed TV productions (raw footage)
- Former client projects (contractually required to keep)
Legacy Systems (5.3 TB):
- Database backups from decommissioned systems
- Old CMS media libraries
- Retired website assets
Unknown Origins (3.2 TB):
- Unlabeled data dumps
- Testing data with no clear ownership
- Personal employee backups (against policy)
đź’ˇ The Breakthrough - Finding Stakeholders
Tracking Down the Owners
For each data category, we needed stakeholder input:
Marketing Team Interview: “Those campaigns are completed. We need to keep them for reference, but we access them maybe once a year at most. We didn’t realize they were costing so much in hot storage.”
Production Team Interview: “That footage is from shows we’ve delivered years ago. We only need access if a client requests reshoots or remasters, which is extremely rare. Archive storage would be fine as long as we can retrieve within 12 hours if needed.”
IT Team Interview: “Those system backups are kept for compliance, but we’ve never needed to restore from them. The retention policy is 7 years, but they definitely don’t need to be in Standard storage.”
⚖️ Risk Assessment - The Migration Strategy
Low-Risk Data Transition
The analysis was clear—this was low-risk, high-reward:
- Data Importance: Required for compliance/contracts, but rarely accessed
- Access Speed Needs: 12-24 hour retrieval acceptable for all identified data
- Workflow Impact: No active projects using this data
- Compliance Requirements: Met by Glacier with 7+ year retention
The Plan of Action
A three-phased approach to migration:
Phase 1: Transition to Glacier (0-30 days)
- Create lifecycle policies for all flagged data
- Implement tiered transition (Standard → IA → Glacier)
- Document all transitioned data with metadata
Phase 2: Policy Implementation (31-60 days)
- Create company-wide storage lifecycle policies
- Implement tagging requirements (project, expiration, owner)
- Build dashboard for storage cost transparency
Phase 3: Long-term Management (61+ days)
- Scheduled quarterly reviews of storage usage
- Automated alerts for potential waste
- Storage cost chargebacks to appropriate teams
🎯 Case Resolution - Executing the Plan
The Migration Operation
The plan was executed flawlessly:
Week 1: Created comprehensive inventory of all 23.2 TB Week 2: Stakeholder sign-offs collected for all data Week 3: Lifecycle policies implemented and transitions begun Week 4: Documentation and training for new storage policies
Immediate Results
- Monthly savings: $8,200
- Annual savings: $98,400
- ROI: 820% (investigation cost vs. annual savings)
- Time to resolution: 10 hours (investigation) + 3 weeks (implementation)
Client Reaction
“I knew we had waste, but I didn’t realize how systematic the problem was. Not only did you find immediate savings, but you’ve given us a framework to prevent future storage sprawl. Our CFO did a literal happy dance when I showed her the numbers.” - Jordan Lee, Director of Cloud Operations
📊 Case Analysis - Lessons Learned
The Hoarding Mentality
Four key factors contributed to this waste:
- “Storage is Cheap” Mindset: While true compared to on-premises, cloud storage still adds up
- Lack of Ownership: No individual accountability for storage costs
- Visibility Gap: Teams unaware of the financial impact of their storage
- Missing Automation: No lifecycle policies to automatically transition data
Prevention Strategy
Implemented for MediaVerse:
- Storage Class Education: Training on appropriate storage tiers
- Required Tagging: Project, owner, expiration date, and access requirements
- Cost Dashboards: Team-level visibility into storage costs
- Default Lifecycle Policies: Automatic transition for all new buckets
- Quarterly Data Cleanup Days: Dedicated time for teams to review and archive
🏆 Case Status: CLOSED
Final Outcome:
- âś… 23.2 TB transitioned to appropriate storage classes
- âś… $98,400 annual savings secured
- âś… Storage governance framework implemented
- âś… Client satisfaction: 10/10
Detective’s Notes: The Storage Hoarder Scandal highlights a common pattern in creative industries—“keep everything, just in case.” This mentality made sense in the early cloud days when storage was perceived as nearly free, but at scale, even small per-gigabyte costs become significant.
What made this case particularly satisfying was that we achieved massive savings without deleting a single file. By matching storage classes to actual access patterns, MediaVerse retained all their valuable assets while dramatically reducing costs.
In cloud storage, it’s not what you keep that costs you—it’s where you keep it.
Need a similar investigation? Contact Detective Cloud Sleuth for your free storage audit.
Tags: S3 Storage Optimization AWS Data Lifecycle Cost Reduction Glacier Media