{"id":8202,"date":"2025-02-20T01:00:00","date_gmt":"2025-02-20T01:00:00","guid":{"rendered":"https:\/\/alphacrc.com\/?post_type=insight&#038;p=8202"},"modified":"2025-04-02T10:06:56","modified_gmt":"2025-04-02T10:06:56","slug":"ai-localization-quantization","status":"publish","type":"insight","link":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/","title":{"rendered":"AI localization: Using quantization to improve adaptability"},"content":{"rendered":"\n<p>With artificial intelligence tools continuing to make inroads into the localization industry, linguistic service providers (LSPs) need to re-evaluate integration methods to reduce turnaround times while preserving translation quality.<\/p>\n\n\n\n<p>Implementing AI localization effectively can be challenging, especially when aiming to do more complex tasks than simply getting an LLM-powered translation. There are often restrictions from popular LLMs sourced through APIs (such as OpenAI) and running larger and more capable open-source models on-site can be expensive.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">AI localization and the open-source problem<\/h2>\n\n\n\n<p>Open-source models can offer businesses a much greater degree of customization and the ability to avoid iterative costs, but they require high-grade hardware in order to run, specifically a good graphics card (GPU) to run calculations, though it is possible, albeit slower, to use the central processing unit (CPU) for this.<\/p>\n\n\n\n<p>Running Llama 3 at full precision means running it at float 32 (FP32), a computer number format that typically occupies 32 bits (four bytes) of memory. The lightest Llama 3 model contains 8 billion parameters, each operating at FP32 precision. Thus:<\/p>\n\n\n\n<p>8,000,000,000 parameters x 4 bytes (memory usage per parameter at FP32) = 32,000,000,000 byte (roughly 30GB) graphics card memory requirement.<\/p>\n\n\n\n<p>But it would be prohibitively expensive to run this Llama 3 on internal hardware. If businesses want to run open-source models such as Llama to perform more varied and nuanced AI experiments, they need to look at quantization techniques.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"622\" src=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-1024x622.png\" alt=\"Group of developers working on AI localization\" class=\"wp-image-8203\" style=\"width:800px;height:auto\" srcset=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-1024x622.png 1024w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-300x182.png 300w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-768x467.png 768w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-1536x934.png 1536w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/AI-dev-2048x1245.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n<h2 class=\"wp-block-heading\">What is quantization?<\/h2>\n\n\n\n<p>In this context, quantization refers to the process of reducing the precision of a model\u2019s weights, which compress the model by reducing the number of bits required to represent each parameter.<\/p>\n\n\n\n<p>Quantization aims to reduce the memory footprint required to run resource-intensive open-source LLMs, even enabling use of edge devices such as phones. Reducing precision also results in faster computation and lower energy consumption.<\/p>\n\n\n\n<p>There are a range of potential quantization methods suitable for AI localization, including post-training quantization, quantization-aware training and dynamic quantization, with the end effect meaning lower requirements to run open-source models.<\/p>\n\n\n\n<p>Going back to Llama 3, quantization offers a reduction in precision from FP32 to FP16 or even float point 8 (FP8). Running Llama 3\u2019s lightest model at FP16 would mean:<\/p>\n\n\n\n<p>&nbsp;8,000,000,000 parameters x 2 bytes (memory usage per parameter at FP16) = 16,000,000,000 byte (roughly 15GB) graphics card memory requirement.<\/p>\n\n\n\n<p>Significant gains such as this can result in massive savings for LSPs looking to experiment with AI integration more freely than is possible through APIs from companies such as OpenAI.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">What are the trade-offs?<\/h2>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>To start with, FP16 has a smaller dynamic range and lower precision compared to FP32, which can result in numerical instability. Gradients could also become less accurate, opening up the potential for slower convergence. This would, of course, be more of an issue during training than inference.<\/p>\n\n\n\n<p>That said, the practical results of this shouldn\u2019t be underestimated. In a translation environment, this could mean subtle nuances or infrequent word associations are lost, or output could become increasingly inaccurate.<\/p>\n\n\n\n<p>Let\u2019s take a look at some potential outcomes of quantization at each step of the LLM-based translation process.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"1001\" src=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-1024x1001.png\" alt=\"Creating AI localization workflows\" class=\"wp-image-8209\" style=\"width:331px;height:auto\" srcset=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-1024x1001.png 1024w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-300x293.png 300w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-768x751.png 768w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-1536x1502.png 1536w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Gears-2048x2002.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h2 class=\"wp-block-heading\">LLM translation and potential quantization issues<\/h2>\n\n\n\n<p>There are four key stages to the LLM-based translation process, each of which could experience issues stemming from quantization if not appropriately addressed.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Input processing<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Tokenization: Input text is broken down into smaller units called tokens.<\/li>\n\n\n\n<li>Encoding: Tokens are converted into numerical representations that the model can process.<\/li>\n<\/ul>\n\n\n\n<p>Potential issues<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Encoding: Quantization can lead to errors in the embedding layer, where continuous space representations of tokens are approximated. This could potentially result in less accurate embeddings, which would limit the model\u2019s understanding of the input text.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Contextual understanding<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Contextual embedding: The model captures the meaning of each token in relation to those surrounding it.<\/li>\n\n\n\n<li>Attention mechanisms: This ensures important information is retained over irrelevant content.<\/li>\n<\/ul>\n\n\n\n<p>Potential issues<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Contextual embedding: Quantization can lower contextual precision, affecting the model\u2019s ability to capture nuances and relationships between tokens.<\/li>\n\n\n\n<li>Attention mechanisms: The model may focus on less relevant parts of the text.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Translation generation<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Decoding: Translation is generated one token at a time.<\/li>\n\n\n\n<li>Beam search: This keeps track of multiple possible translations and selects the best one based on a scoring mechanism.<\/li>\n<\/ul>\n\n\n\n<p>Potential issues<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Decoding: Precision of generated tokens may be lower. The model may therefore generate tokens that are close but not exactly correct.<\/li>\n\n\n\n<li>Beam search: The scoring mechanism may be affected, leading to suboptimal translation path selection.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Output processing<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Detokenization: Generated tokens are converted back into human-readable text.<\/li>\n\n\n\n<li>Post-processing: Additional steps may be taken to improve the translation, such as correcting grammar.<\/li>\n<\/ul>\n\n\n\n<p>Potential issues<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Post-processing: As a step in itself, post-processing will be less affected by quantization, but any errors introduced in previous stages can propagate and would become more noticeable here.<\/li>\n<\/ul>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance-1024x576.png\" alt=\"Working on effective AI localization can be like putting the pieces of a jigsaw together.\" class=\"wp-image-8214\" srcset=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance-1024x576.png 1024w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance-300x169.png 300w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance-768x432.png 768w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance-1536x864.png 1536w, https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/Puzzle-dance.png 1920w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h2 class=\"wp-block-heading\">Mitigating the effects of quantization<\/h2>\n\n\n\n<p>For AI localization, it\u2019s vital that accuracy and context need to be preserved, meaning that the potential impacts of quantization must be mitigated. There are several methods that could aid with this.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Mixed-precision quantization<\/h3>\n\n\n\n<p>While quantization can significantly reduce hardware requirements, it doesn\u2019t need to be applied across the entire model. Mixed-precision quantization keeps critical layers at higher prevision levels such as FP32, while reducing other layers to lower points such as FP16.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Quantization-aware training<\/h3>\n\n\n\n<p>Businesses can also consider training the model with quantization in mind. By simulating the effects of quantization during training, the model can learn how best to compensate for the lower precision levels. This typically results in better performance when compared to processes such as raw post-training quantization.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Post-quantization fine-tuning<\/h3>\n\n\n\n<p>By fine-tuning on a relevant dataset after the quantization process is complete, the model can better adjust to its lower precision and recover some of its lost performance.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Error-feedback mechanisms<\/h3>\n\n\n\n<p>Implementing error-feedback mechanisms, such as the employment of human specialists can help the model to correct quantization errors during the inference stage of translation. The model can maintain the quality of translated or localized content from feedback generated over time.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link has-background wp-element-button\" href=\"https:\/\/alphacrc.com\/llm-datasets\/\" style=\"background-color:#fc3065\">Need help with LLM datasets? Alpha CRC is ready to help.<\/a><\/div>\n<\/div>\n\n\n\n<h2 class=\"wp-block-heading\">Potential applications for quantization in AI localization<\/h2>\n\n\n\n<p>LSPs have been exploring potential applications of AI since 2022, but hardware requirements and specific regulations present issues using APIs from the likes of OpenAI. Quantization techniques go some way to help, if applied with care and any negative impacts can be mitigated through the approaches stated above.<\/p>\n\n\n\n<p>A quantized model can help reduce internal content production and adaptation costs, as access to open-source models would eliminate the need to use APIs from major LLM providers, ultimately resulting in a more widely-adopted AI localization workflow. Lower hardware requirements also mean that the technology could be applied more readily throughout a company\u2019s workflows.<\/p>\n\n\n\n<p>Open-source models are more flexible than those provided through API access, offering more potential for fine-tuning; the ability to run open-source models from internal hardware enhances security protocols and reduces the risks of IP leakage.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">About Alpha CRC<\/h2>\n\n\n\n<p>Alpha CRC offers clients high-quality localization services that blend the best of human creativity with the speed and power of technology.<\/p>\n\n\n\n<p>From translation to content creation, Alpha CRC enables clients to engage with their global customer bases and improve their reach.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link has-background wp-element-button\" href=\"https:\/\/alphacrc.com\/llm-fine-tuning\/\" style=\"background-color:#fc3065\">Find out more about Alpha CRC&rsquo;s LLM fine-tuning services<\/a><\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>With artificial intelligence tools continuing to make inroads into the localization industry, linguistic service providers (LSPs) need to re-evaluate integration methods to reduce turnaround times while preserving translation quality. Implementing AI localization effectively can be challenging, especially when aiming to do more complex tasks than simply getting an LLM-powered translation. There are often restrictions from [&hellip;]<\/p>\n","protected":false},"author":14,"featured_media":8223,"template":"","service":[],"sector":[],"insight_type":[],"class_list":["post-8202","insight","type-insight","status-publish","has-post-thumbnail","hentry"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v19.1 (Yoast SEO v24.2) - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>AI localization using quantization - Alpha CRC<\/title>\n<meta name=\"description\" content=\"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the &#039;somewhere-in-the-middles&#039;.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/\" \/>\n<meta property=\"og:locale\" content=\"fr_FR\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"AI localization: Using quantization to improve adaptability\" \/>\n<meta property=\"og:description\" content=\"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the &#039;somewhere-in-the-middles&#039;.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/\" \/>\n<meta property=\"og:site_name\" content=\"Alpha CRC\" \/>\n<meta property=\"article:modified_time\" content=\"2025-04-02T10:06:56+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"2560\" \/>\n\t<meta property=\"og:image:height\" content=\"1707\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Dur\u00e9e de lecture estim\u00e9e\" \/>\n\t<meta name=\"twitter:data1\" content=\"7 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/\",\"url\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/\",\"name\":\"AI localization using quantization - Alpha CRC\",\"isPartOf\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg\",\"datePublished\":\"2025-02-20T01:00:00+00:00\",\"dateModified\":\"2025-04-02T10:06:56+00:00\",\"description\":\"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the 'somewhere-in-the-middles'.\",\"breadcrumb\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#breadcrumb\"},\"inLanguage\":\"fr-FR\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"fr-FR\",\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage\",\"url\":\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg\",\"contentUrl\":\"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg\",\"width\":2560,\"height\":1707,\"caption\":\"Developers working on AI localization\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/alphacrc.com\/fr\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"AI localization: Using quantization to improve adaptability\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/alphacrc.com\/fr\/#website\",\"url\":\"https:\/\/alphacrc.com\/fr\/\",\"name\":\"Alpha CRC\",\"description\":\"Bespoke localization, translation and more since 1987\",\"publisher\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/alphacrc.com\/fr\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"fr-FR\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/alphacrc.com\/fr\/#organization\",\"name\":\"Alpha CRC\",\"url\":\"https:\/\/alphacrc.com\/fr\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"fr-FR\",\"@id\":\"https:\/\/alphacrc.com\/fr\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/alphacrc.com\/wp-content\/uploads\/2022\/02\/logo.png\",\"contentUrl\":\"https:\/\/alphacrc.com\/wp-content\/uploads\/2022\/02\/logo.png\",\"width\":184,\"height\":184,\"caption\":\"Alpha CRC\"},\"image\":{\"@id\":\"https:\/\/alphacrc.com\/fr\/#\/schema\/logo\/image\/\"}}]}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"AI localization using quantization - Alpha CRC","description":"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the 'somewhere-in-the-middles'.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/","og_locale":"fr_FR","og_type":"article","og_title":"AI localization: Using quantization to improve adaptability","og_description":"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the 'somewhere-in-the-middles'.","og_url":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/","og_site_name":"Alpha CRC","article_modified_time":"2025-04-02T10:06:56+00:00","og_image":[{"width":2560,"height":1707,"url":"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg","type":"image\/jpeg"}],"twitter_card":"summary_large_image","twitter_misc":{"Dur\u00e9e de lecture estim\u00e9e":"7 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/","url":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/","name":"AI localization using quantization - Alpha CRC","isPartOf":{"@id":"https:\/\/alphacrc.com\/fr\/#website"},"primaryImageOfPage":{"@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage"},"image":{"@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage"},"thumbnailUrl":"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg","datePublished":"2025-02-20T01:00:00+00:00","dateModified":"2025-04-02T10:06:56+00:00","description":"Find out more about the implications of a quantized LLM for AI localization - the good, the bad, and the 'somewhere-in-the-middles'.","breadcrumb":{"@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#breadcrumb"},"inLanguage":"fr-FR","potentialAction":[{"@type":"ReadAction","target":["https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/"]}]},{"@type":"ImageObject","inLanguage":"fr-FR","@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#primaryimage","url":"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg","contentUrl":"https:\/\/alphacrc.com\/wp-content\/uploads\/2025\/02\/creative-asian-man-with-headphones-using-laptop-at-2024-03-20-16-31-36-utc-scaled.jpg","width":2560,"height":1707,"caption":"Developers working on AI localization"},{"@type":"BreadcrumbList","@id":"https:\/\/alphacrc.com\/fr\/insight\/ai-localization-quantization\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/alphacrc.com\/fr\/"},{"@type":"ListItem","position":2,"name":"AI localization: Using quantization to improve adaptability"}]},{"@type":"WebSite","@id":"https:\/\/alphacrc.com\/fr\/#website","url":"https:\/\/alphacrc.com\/fr\/","name":"Alpha CRC","description":"Bespoke localization, translation and more since 1987","publisher":{"@id":"https:\/\/alphacrc.com\/fr\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/alphacrc.com\/fr\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"fr-FR"},{"@type":"Organization","@id":"https:\/\/alphacrc.com\/fr\/#organization","name":"Alpha CRC","url":"https:\/\/alphacrc.com\/fr\/","logo":{"@type":"ImageObject","inLanguage":"fr-FR","@id":"https:\/\/alphacrc.com\/fr\/#\/schema\/logo\/image\/","url":"https:\/\/alphacrc.com\/wp-content\/uploads\/2022\/02\/logo.png","contentUrl":"https:\/\/alphacrc.com\/wp-content\/uploads\/2022\/02\/logo.png","width":184,"height":184,"caption":"Alpha CRC"},"image":{"@id":"https:\/\/alphacrc.com\/fr\/#\/schema\/logo\/image\/"}}]}},"_links":{"self":[{"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/insight\/8202","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/insight"}],"about":[{"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/types\/insight"}],"author":[{"embeddable":true,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/users\/14"}],"version-history":[{"count":9,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/insight\/8202\/revisions"}],"predecessor-version":[{"id":8541,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/insight\/8202\/revisions\/8541"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/media\/8223"}],"wp:attachment":[{"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/media?parent=8202"}],"wp:term":[{"taxonomy":"service","embeddable":true,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/service?post=8202"},{"taxonomy":"sector","embeddable":true,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/sector?post=8202"},{"taxonomy":"insight_type","embeddable":true,"href":"https:\/\/alphacrc.com\/fr\/wp-json\/wp\/v2\/insight_type?post=8202"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}