透過集合功能整理內容
你可以依據偏好儲存及分類內容。
Google 如何檢索會自動調整語言代碼的網頁
如果您網站中的網頁會「自動調整語言代碼」(亦即網站會判斷訪客所在的國家/地區或語言偏好,並依此傳回不同內容),那麼 Google 可能不會為所有語言代碼的內容進行檢索、排名或建立索引。這是因為 Googlebot 檢索器的預設 IP 位址會顯示為位於美國,而且檢索器傳送 HTTP 要求時也不會在要求標頭中設定 Accept-Language
。
地理區域分散式檢索作業
除了美國 IP 位址之外,Googlebot 也會使用位於美國以外地區的 IP 位址進行檢索。
我們向來推薦的做法是,無論 Googlebot 顯示為來自哪個國家/地區,都將 Googlebot 視為該國家/地區的一般使用者處理。也就是說,如果您禁止位於美國的使用者存取內容,但允許位於澳洲的訪客查看內容,那麼您的伺服器應該封鎖具有美國 IP 位址的 Googlebot,同時允許具有澳洲 IP 位址的 Googlebot 存取內容。
其他注意事項
除非另有註明,否則本頁面中的內容是採用創用 CC 姓名標示 4.0 授權,程式碼範例則為阿帕契 2.0 授權。詳情請參閱《Google Developers 網站政策》。Java 是 Oracle 和/或其關聯企業的註冊商標。
上次更新時間:2025-08-04 (世界標準時間)。
[null,null,["上次更新時間:2025-08-04 (世界標準時間)。"],[[["\u003cp\u003eGooglebot's default IP addresses appear to be US-based, which may impact the crawling and indexing of locale-adaptive pages targeting other regions.\u003c/p\u003e\n"],["\u003cp\u003eIt is recommended to use separate locale URL configurations with \u003ccode\u003erel="alternate"\u003c/code\u003e hreflang annotations for better localization.\u003c/p\u003e\n"],["\u003cp\u003eGooglebot crawls from various global locations, so treat it like any other user based on its apparent location, including access restrictions.\u003c/p\u003e\n"],["\u003cp\u003eEnsure consistent robots exclusion protocol (robots.txt and meta tags) across all locales to avoid unintended crawling restrictions.\u003c/p\u003e\n"]]],["Google crawls locale-adaptive pages using IP addresses from various locations, not just the USA. When Googlebot appears to be from a specific country, treat it like a user from that region. For locale-adaptive sites, using separate URL configurations with `rel=\"alternate\"` hreflang annotations is recommended. Ensure consistent application of robots exclusion protocols, such as robots.txt and meta tags, across all locales. You can verify Googlebot's geo-distributed crawls through reverse DNS lookups.\n"],null,["# How Google Crawls Locale-Adaptive Pages | Google Search Central\n\nHow Google crawls locale-adaptive pages\n=======================================\n\nIf your site has *locale-adaptive* pages (that is, your site returns different content\nbased on the perceived country or preferred language of the visitor), Google might not crawl,\nindex, or rank all your content for different locales. This is because the default IP\naddresses of the Googlebot crawler appear to be based in the USA. In addition, the crawler\nsends HTTP requests without setting `Accept-Language` in the request header.\n| **Important** : We recommend using separate locale URL configurations and annotating them with [`rel=\"alternate\"`\n| hreflang annotations](/search/docs/specialty/international/localized-versions).\n\nGeo-distributed crawling\n------------------------\n\nGooglebot crawls with IP addresses based outside the USA, in addition to the US-based IP addresses.\n\nAs we have always recommended, when Googlebot appears to come from a certain country, treat\nit like you would treat any other user from that country. This means that if you block\nUSA-based users from accessing your content, but allow visitors from Australia to see it,\nyour server should block Googlebot if it appears to be coming from the USA, but allow access\nto Googlebot if it appears to come from Australia.\n\n### Other considerations\n\n- Googlebot uses the same user agent string for all crawling configurations. Learn more about the [user agent strings used\n by Google crawlers](/search/docs/crawling-indexing/overview-google-crawlers).\n- You can [verify Googlebot\n geo-distributed crawls](/search/docs/crawling-indexing/verifying-googlebot) using reverse DNS lookups.\n- If your site is using the [*robots exclusion protocol*](https://www.rfc-editor.org/rfc/rfc9309.html), make sure you apply it consistently across locales. This means that [robots `meta` tags](/search/docs/crawling-indexing/robots-meta-tag) and the [robots.txt file](/search/docs/crawling-indexing/robots/create-robots-txt) must specify the same rules in each locale."]]