Mobile proxies for ecommerce price intelligence 2026
Mobile proxies for ecommerce price intelligence 2026
ecommerce price intelligence is the practice of monitoring competitor prices, stock levels, promotions, and product assortment across marketplaces and brand sites at high frequency. it is foundational for any serious retailer, brand, or marketplace operator. in 2026, the work is harder than it was three years ago because anti-bot detection has matured aggressively, especially on the largest marketplaces (Amazon, Shopee, Lazada, Tokopedia, JD, Mercado Libre). mobile proxies have become the operational baseline for high-volume price intelligence pipelines because their detection profile beats residential and ISP IPs on the platforms that matter most.
this post covers how to build a price intelligence stack that works in 2026: the workflow shape, the proxy rotation strategy, the anti-bot considerations, and the specific tactics for Singapore marketplaces where SingaporeMobileProxy is purpose-built for the work. read on for the technical detail and a honest comparison of when mobile proxies are the right buy versus when residential or ISP suffice.
what price intelligence actually requires
a working price intelligence pipeline does five things, repeatedly, at scale.
- product discovery: maintain a current catalog of competitor SKUs across categories, with mapping back to your own SKUs.
- price scraping: collect current price, list price, sale price, and promotion details for each tracked SKU at intervals from hourly to weekly depending on category volatility.
- stock and availability tracking: detect stockouts, “low stock” warnings, shipping windows, and restocking patterns.
- promotion and bundle detection: identify flash sales, BOGO offers, percentage discounts, voucher stacks, and free shipping thresholds.
- assortment monitoring: detect new product launches, discontinuations, and category expansions.
each of these requires hundreds to millions of HTTP requests per day depending on category breadth. each request is to a major marketplace whose anti-bot stack is sophisticated. detection failures lead to one of three bad outcomes: blocked IPs (waste budget), captcha walls (waste engineering time), or stale data (waste decisions). the IP infrastructure choice determines which of these happens.
the proxy choice for price intelligence
| proxy type | success rate on top marketplaces | cost profile | best fit |
|---|---|---|---|
| datacenter | low to medium, often blocked | very low cost | internal QA, low-volume tracking |
| residential | medium to high | medium-high cost per GB | most public marketplace scraping |
| ISP (static residential) | medium to high | medium cost per GB | sticky session scraping |
| dedicated mobile | high | per-port flat fee | logged-in scraping, mobile-first marketplaces |
the high-volume reality: most price intelligence runs on residential or ISP proxies because they balance cost and success rate well. for Singapore marketplaces (Shopee SG, Lazada SG, Carousell), where mobile-first apps dominate consumer behavior, mobile proxies meaningfully outperform residential because the marketplace anti-bot fingerprints expect mobile-carrier ASN traffic.
our TikTok Shopee Lazada Singapore guide goes deep on the platform-specific detection patterns. for Singapore-focused price intelligence on Shopee, Lazada, and Carousell, dedicated mobile is the operational baseline.
reference architecture
a working pipeline looks roughly like this:
[ catalog DB ] --> [ scheduler ] --> [ scraping workers ]
|
v
[ proxy pool ]
|
v
[ marketplace ]
|
v
[ raw response store ]
|
v
[ parser + diff engine ]
|
v
[ price warehouse ]
|
v
[ dashboards + alerts ]
each component has a specific role. the scheduler queues SKUs by category cadence (electronics may need hourly checks during flash sales, household goods may need weekly). workers pull SKUs from the queue and fetch product pages through the proxy pool. raw responses go to durable storage (S3, GCS) before parsing because anti-bot signatures change frequently and you want to be able to re-parse historical responses without re-scraping. the diff engine compares current prices against historical and emits alerts for material changes. the warehouse holds the long-term data for dashboards, BI, and ML pricing models.
sample scraper with smp
import requests
import time
import random
PROXY = {
"http": "http://user_xyz:pass_abc@sg-mobile.singaporemobileproxy.com:10001",
"https": "http://user_xyz:pass_abc@sg-mobile.singaporemobileproxy.com:10001",
}
HEADERS_MOBILE = {
"User-Agent": "Mozilla/5.0 (Linux; Android 13; SM-S918B) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Mobile Safari/537.36",
"Accept-Language": "en-SG,en;q=0.9",
}
def fetch_product(url):
resp = requests.get(url, proxies=PROXY, headers=HEADERS_MOBILE, timeout=30)
if resp.status_code in (429, 403):
rotate_proxy()
time.sleep(random.uniform(5, 15))
return None
return resp.text
def rotate_proxy():
requests.post(
"https://singaporemobileproxy.com/rotate/<your_token>",
headers={"X-API-Key": "<your_api_key>"},
timeout=10,
)
def scrape_batch(urls):
results = []
for url in urls:
for attempt in range(3):
html = fetch_product(url)
if html:
results.append((url, html))
break
time.sleep(2 ** attempt)
return results
this is the rough shape. production code adds retry queues, circuit breakers per host, response deduplication, parser version tracking, and observability hooks. the proxy layer is the thin part: a simple HTTP proxy with rotation triggered on bad status codes.
rotation strategy
three rotation patterns work for price intelligence:
fixed-interval rotation: rotate every N minutes regardless of activity. simple, predictable. wastes some IP freshness but works well for high-volume scraping where you do not need session continuity per SKU.
bad-status rotation: rotate when the current IP gets a 429, 403, or captcha. preserves IP freshness for as long as possible. requires good circuit breaker logic to avoid hammering a flagged IP.
session-aware rotation: rotate at the boundary between scraping sessions (e.g., when changing categories or workers). good fit for workloads that need short-lived stable sessions.
for Shopee SG and Lazada SG, fixed-interval rotation every 15 to 30 minutes works well combined with bad-status rotation as a fallback. for Carousell, longer intervals (1 to 2 hours) work because the platform is less aggressive on rate limiting.
handling anti-bot defenses
major marketplaces deploy multiple anti-bot layers. understanding each layer is necessary for designing a scraper that does not get blocked.
layer 1: IP reputation. datacenter IPs get scored down or blocked at the first request. residential IPs pass unless flagged for prior bot activity. mobile IPs share CGNAT with thousands of real users and are very hard to blocklist without collateral damage. mitigation: use mobile or residential.
layer 2: TLS and HTTP fingerprinting. requests from Python’s default requests library have distinctive TLS fingerprints (JA3 hashes) that differ from real browsers. some marketplaces blocklist non-browser TLS signatures. mitigation: use curl_cffi or tls_client libraries that mimic browser TLS, or use a real browser via Playwright or Puppeteer.
layer 3: behavioral analysis. rapid sequential requests, no mouse movement, no scrolls, no realistic dwell times. modern anti-bot tracks request patterns over a session. mitigation: pace requests, randomize timing, add realistic browsing patterns when using real browsers.
layer 4: cookie and session state. many marketplaces require a warm session (set cookies, accept tracking) before serving full product data. cold requests get partial data or redirects. mitigation: maintain session state per IP, warm up the session before scraping.
layer 5: device fingerprinting. canvas fingerprint, WebGL fingerprint, font list, screen resolution, audio context, hardware concurrency. visible only to JS execution. mitigation: use a real browser with antidetect features, rotate device fingerprints alongside IP.
mobile proxies primarily address layer 1. layers 2 through 5 require additional engineering. the success of a price intelligence pipeline comes from doing all five layers reasonably well, not from any single layer alone. our web scraping with mobile proxies guide goes deeper on the layered approach.
singapore-specific tactics
scraping Shopee SG, Lazada SG, and Carousell from a Singapore mobile carrier IP gives you a meaningfully different experience than from a residential or datacenter IP. the marketplace serves Singapore-localized product listings, prices in SGD, shipping options to Singapore addresses, and Singapore-specific promotions. the JS bundles include Singapore market features. the anti-bot scoring is calibrated to expect mobile-carrier traffic from local users.
practical impact: a single SMP port (SGD 60/month) scraping Shopee SG, Lazada SG, and Carousell at moderate frequency typically delivers 90 to 95 percent success rate without aggressive captcha walls. the same volume from a residential pool typically delivers 75 to 85 percent. the difference compounds over millions of requests per month and shows up in cleaner data, fewer reprocessing cycles, and faster end-to-end latency.
diff engine and alerting
once the raw data lands, the diff engine compares each new observation against historical. material changes trigger alerts:
- price drop more than 5 percent: alert pricing team, evaluate competitive response.
- stockout transition: alert merchandising, validate own stock position.
- new SKU launched in tracked category: alert product team.
- bundle or promotion change: alert marketing.
- shipping fee change: alert logistics.
alert routing depends on team structure. modern stacks send to Slack, Teams, or email with summary tables and links to the source product pages. the dashboards layer holds the historical view: price curves over 30/90/365 days, stockout frequency by SKU, promotion calendar overlays.
storage and re-scrapability
the most important architectural decision is to store raw HTML responses (or full DOM snapshots) durably. anti-bot signatures change, marketplaces redesign category pages, and your parsers will fail occasionally. when they fail, you want to be able to re-parse historical responses without re-scraping (which costs proxy bandwidth and risks triggering rate limits on the marketplace). store responses in S3 or GCS with content-addressed paths and metadata indexing for retrieval.
a typical schema:
s3://price-intel/
YYYY/MM/DD/HH/
{marketplace}/
{sku_hash}.html.gz
index/
YYYY-MM-DD.parquet (sku, marketplace, fetched_at, content_hash, status_code)
the parquet index lets you query historical fetches by SKU and marketplace without listing object storage. the gzipped HTML keeps storage cost modest.
working with brand pages versus marketplaces
brand pages (manufacturer’s own sites) are typically easier to scrape than marketplaces because their anti-bot is less aggressive. residential or ISP proxies are usually sufficient. allocate dedicated mobile budget to the marketplaces (Shopee, Lazada, Amazon, etc.) and use cheaper IPs for brand pages.
for Singapore-specific brand pages, mobile IPs still help because the brand site may serve different content to mobile-carrier users (mobile-optimized layout, app download prompts, Singapore-specific contact info). but the success rate on residential is high enough that mobile is rarely necessary for brand sites.
comparison: mobile vs alternatives for price intelligence
| dimension | mobile (SMP) | residential | ISP | datacenter |
|---|---|---|---|---|
| Shopee SG success rate | 95 percent plus | 80 to 90 percent | 80 to 90 percent | flaky |
| Lazada SG success rate | 95 percent plus | 80 to 90 percent | 80 to 90 percent | flaky |
| Amazon US success rate | 90 to 95 percent | 85 to 95 percent | 85 to 95 percent | low |
| cost per million requests | high (volume-independent) | medium-high | medium | low |
| best for | high-detection marketplaces | mid-detection marketplaces | sticky sessions | low-detection sites |
for related deep-dives, see our forthcoming mobile proxies for SERP scraping, mobile proxies for brand monitoring, and mobile proxies for travel and flight scraping posts in this same wave.
faq
is scraping marketplace prices legal?
scraping public product pages is generally legal in most jurisdictions including Singapore (PDPA does not cover product information that is not personal data). marketplace terms of service may prohibit scraping, but ToS violations are typically civil (potentially actionable by the marketplace) not criminal. consult your legal team for jurisdiction-specific guidance.
how often should I rescrape a SKU?
depends on category. fashion and electronics during flash sales: every 15 to 60 minutes. household goods: daily. books and slow-moving inventory: weekly. set cadence per category and adjust based on observed price volatility.
do I need a real browser or can I use HTTP requests?
HTTP requests work for many marketplaces if you handle TLS fingerprinting, headers, and cookies correctly. for the most aggressive anti-bot (Amazon, some Shopify storefronts), real browsers via Playwright are more reliable. budget engineering accordingly.
what is the cost of a working price intelligence stack?
for Singapore marketplace coverage at moderate volume (hundreds of thousands of requests per month): SGD 200 to 500 per month for proxies, SGD 100 to 300 for storage, plus engineering time. for global multi-marketplace coverage: USD 5,000 to 50,000 per month for proxies depending on volume.
can I run price intelligence on a single SMP port?
yes for low-volume Singapore-only scraping (a few thousand SKUs at daily cadence). for higher volume or multiple categories, scale up by buying additional ports. each port handles its own concurrency limits.
start with a 2-hour trial
we offer a 2-hour free trial with no credit card required. dedicated SingTel, StarHub, or M1 modem, real Singapore mobile IP, instant rotation, HTTP and SOCKS5. point your scraper at Shopee SG, Lazada SG, or Carousell, run for two hours, and verify the success rate against your existing residential or datacenter pipeline.