Batch geocoding converts a list of street addresses into latitude and longitude coordinates in a single operation. Instead of looking up one address at a time, you send hundreds or thousands of records through a geocoding service and get structured coordinates back for each one.
Organizations that manage large address datasets, from bank loan portfolios to insurance policy files to retail store networks, rely on batch geocoding to turn raw addresses into mappable, analyzable location data. The alternative is manual lookups, which breaks down the moment your dataset crosses a few dozen records.
Below, you’ll learn how to prepare your data, pick the right batch geocoding method, and validate results at scale.
Try a single lookup below — for batch processing, use our API.
When You Need Batch Geocoding vs Single Lookups
A single geocoding request works fine when you need coordinates for one address. You enter it, you get a result. But once your dataset has more than about 100 records, single lookups become a bottleneck.
Batch geocoding solves three problems at once. First, speed: a batch API endpoint can process 5,000 addresses in the time it takes to manually look up 10. Second, cost: most geocoding providers charge less per record at batch volumes than per individual request. Third, consistency: every address runs through the same parsing and matching logic, so your output quality stays uniform.
If you need to geocode a single address, start there. For everything else, batch is the way to go.
How to Prepare Your Data for Batch Geocoding
The biggest factor in batch geocoding success is data quality going in. Garbage addresses produce garbage coordinates. Spending 30 minutes cleaning your dataset before upload can push your match rate from 80% to 95%+.
Standardize Address Formats
Every address in your dataset should follow the same structure. Separate street, city, state, and postal code into individual columns rather than cramming everything into a single field. For international datasets, add a country code column (ISO 3166-1 alpha-2 works best).
The USPS Publication 28 defines the standard format for U.S. addresses. Following it catches common errors like misabbreviated street types (St vs Street vs Str) before they reach the geocoder.
Clean and Validate Before Upload
Remove duplicate rows. Fix obvious typos in city names. Fill in missing postal codes where possible. Strip special characters, extra spaces, and line breaks from address fields. Each of these small fixes adds up: a clean dataset typically achieves 5-10 percentage points higher match rate than a raw export.
Choose the Right File Format
Most batch geocoding services accept CSV files, and for good reason. CSV is lightweight, universally supported, and easy to inspect in any spreadsheet application. JSON works better for API-to-API workflows where you’re programmatically building the request. Excel files (.xlsx) are convenient for teams that work in spreadsheets, though some services require conversion to CSV first.
Whichever format you pick, map your columns clearly. Label them street, city, state, postal_code, and country so the geocoder knows which field is which.
Batch Geocoding Methods Compared
Not every batch geocoding approach fits every situation. The right choice depends on your volume, budget, and whether you need a one-time conversion or an ongoing pipeline.
| Method | Speed | Scale | Cost | Accuracy | Best For |
|---|---|---|---|---|---|
| API batch endpoint | Fast | 5,000+ per call | Pay-per-use | High | Production pipelines |
| Desktop GIS software | Moderate | Thousands | License fee | High | One-time projects |
| Free online tools | Slow | Hundreds | Free | Variable | Small datasets |
| Spreadsheet plugins | Slow | Hundreds | Free/cheap | Variable | Ad hoc analysis |
API batch endpoints are the standard for production workflows. You send a JSON or CSV payload, the service processes all addresses asynchronously, and you retrieve results when the job completes. Most geocoding APIs offer dedicated batch endpoints with higher rate limits than their single-address counterparts.
Desktop GIS software like ArcGIS includes built-in batch geocoding tools. These work well for GIS analysts who already have a license, but they require local processing power and don’t scale as smoothly for recurring jobs.
Free online tools and spreadsheet plugins handle small datasets without any cost. The Census Bureau geocoder, for example, accepts up to 10,000 U.S. addresses per batch. The tradeoff is speed (typically 1-2 requests per second) and limited international coverage.

Quality Metrics That Matter
A batch geocoding job isn’t done when the coordinates come back. You need to verify that the results are actually usable. Three metrics tell you whether your batch succeeded or needs rework.
Match rate measures what percentage of your input addresses returned a coordinate. A well-prepared dataset should hit 95% or higher. Anything below 90% usually signals formatting issues in the input data rather than problems with the geocoder itself.
Accuracy grade tells you how precisely each coordinate was placed. Rooftop-level matches pin the exact building. Interpolated matches estimate the position along a street segment. Centroid matches drop the pin at the center of a postal code or city, which may be miles from the actual address. Continuuiti’s geocoder scores every result with accuracy grades, so you can filter low-confidence matches before they reach your downstream systems.
Confidence score reflects how certain the geocoder is about each match. Low-confidence results often mean the input was ambiguous (e.g., “123 Main St” without a city or state). Flag these for manual review rather than accepting them blindly.

Scaling Batch Geocoding to Millions of Records
Processing a few thousand addresses is straightforward. Scaling to hundreds of thousands or millions introduces new challenges around rate limits, processing time, and cost management.
Partition by region. Split your dataset by country or state before sending it through the geocoder. Regional partitions run faster because the geocoding engine searches a smaller reference dataset. They also let you route different regions to different providers if one has better coverage for specific areas.
Use asynchronous processing. For large jobs, submit the batch and poll for completion rather than waiting on a synchronous response. Most batch geocoding APIs return a job ID that you check periodically. The job runs server-side, and you retrieve results when it finishes.
Cache results. If you’re geocoding the same addresses repeatedly (e.g., re-running monthly loan portfolios), cache previous results and only geocode new or changed records. Caching can cut your geocoding volume by 60-80% on recurring datasets.
Organizations running climate vulnerability assessments across thousands of facility locations often batch geocode their entire portfolio first, then pipe the coordinates into hazard and risk models. Getting the geocoding step right saves time and cost in every downstream analysis.
Free vs Paid Batch Geocoding Options
Free batch geocoding tools exist, but they come with limits that matter at scale. Here’s how they compare to commercial options.
| Feature | Free (Nominatim, Census Bureau) | Paid (Commercial API) |
|---|---|---|
| Rate limits | 1-2 req/sec | 50,000+/day |
| Batch endpoint | No (loop required) | Yes |
| Quality scoring | No | Yes |
| Global coverage | Variable (Census = US only) | 190+ countries |
| SLA / Support | None | Available |
Free tools like Nominatim (powered by OpenStreetMap) and the Census Bureau geocoder work for small, U.S.-focused projects. Nominatim caps usage at roughly 1 request per second and prohibits bulk downloads. The Census geocoder handles up to 10,000 addresses per file but only covers U.S. locations.
Paid batch geocoding software and API services remove these constraints. Commercial providers offer true batch endpoints, per-record quality scoring, international address coverage, and volume pricing that drops the per-address cost as your usage grows. For recurring workflows or datasets above 10,000 records, the cost usually pays for itself in time savings alone.

FAQ
What is batch geocoding?
Batch geocoding is the process of converting a large list of addresses into latitude and longitude coordinates in a single operation. Instead of geocoding one address at a time, you submit hundreds or thousands of records and receive coordinates for all of them at once.
How many addresses can you geocode at once?
It depends on the provider. Free tools like the Census Bureau geocoder accept up to 10,000 addresses per file. Commercial API batch endpoints typically handle 5,000 to 50,000 addresses per request, with no practical limit on total volume across multiple requests.
What file format works best for batch geocoding?
CSV is the most widely supported format for batch geocoding. It’s lightweight, easy to inspect, and accepted by nearly every geocoding service. JSON works better for programmatic API-to-API workflows. Label your columns clearly (street, city, state, postal_code, country) for the best results.
What is a good geocoding match rate?
A well-prepared dataset should achieve a 95% or higher match rate. Anything below 90% typically indicates formatting issues in the input data, such as missing postal codes, inconsistent abbreviations, or addresses crammed into a single field.
Is batch geocoding free?
Some batch geocoding tools are free but come with limitations. The Census Bureau geocoder is free for U.S. addresses (up to 10,000 per file). Nominatim is free but caps usage at about 1 request per second. For larger datasets, international coverage, or quality scoring, commercial providers offer paid batch geocoding with volume discounts.
How long does it take to geocode 10,000 addresses?
With a commercial batch API endpoint, 10,000 addresses typically process in under 60 seconds. Free tools are significantly slower due to rate limits. Using Nominatim at 1 request per second, 10,000 addresses would take roughly 3 hours. The Census Bureau batch endpoint processes 10,000 in a few minutes.
Batch geocoding turns raw address lists into structured, coordinate-enriched datasets that feed into mapping, analysis, and risk models. The key to a successful batch run is clean input data, the right method for your volume, and a validation step that catches low-quality matches before they propagate downstream. Start with your data preparation, pick a method that fits your scale, and always check your match rate before moving on.
