Localization, Localisation

Practical and concise answers to common questions in G11N, I18N and L10N

Posts Tagged ‘L10N’

Rookie Story: Where to Start with Localisation Management?

Posted by Nick Peris on October 11, 2011

Congratulations! You aced that interview a few weeks ago, and this morning you strolled into the office with a spring in your step! You had the HR induction and were introduced to your new colleagues. Now you’re logging onto the network, the company handbook reassuringly lying on the corner of your desk, or saved on your desktop.

Time to get started! The Company hired you to bring under control this thing almost mysteriously referred to as “Translations”. Your objectives are simple: reduce cost and improve quality. You are their first ever Localisation Manager, and you know the keys to your success will be the   standardisation and centralisation of all Localisation activities.

So what do you need to consider from a technical and organisational point of view?Flags, Nations, People

Getting to Know your Internal Customers

If there have been Translations in your Organisation, there are existing processes and linguistic assets you should be able to build on. You need to quickly learn about them by focussing on:

  1. Who are your allies? Each Department, Local Office etc. probably has at least one “Translation person”. Find out who they are and what they have been doing. Determine whether they will remain involved once you’ve established the new structure, or if they expect to be relieved of Localisation duties. All going well, you may be able to enroll some of them in an inter-departmental Localisation team, even if it’s only a virtual team.
  2. What is the inventory of current processes? Meet the current owners and document everything. No need for anything fancy since you are going to change these processes, but you need to have it all down so that when the inventory is finished you have an accurate and complete picture.
  3. What are the points common to all? Which of those processes work well and which don’t? The successful ones will be the building blocks for your future world.
  4. What are the specificities of each one? Which are worth keeping? Can they be used by other parts of the Organisation? Do they need to remain specific? Your new processes will need to achieve a balance between harmonisation and flexibility.
  5. Do any of those existing processes use technology such as CAT Tools, Content Management Systems, Translation Management Systems? If so should they be upscaled and shared across the Organization?
  6. Do any maintain linguistic assets like Glossaries, Style guides, Translation Memories or even just bilingual files which could be used to create TM’s?

Understanding your product lines

You need to understand what you are going to localise thoroughly before you can develop the processes. The question to answer are:

  1. What types of content: marketing, commercial website, Software, Help systems, self-service technical content, user-driven content like blogs etc. all those use very different registers, vocabulary, address etc. Moreover the choices made will differ again from one language to the next. Some content types require high volumes at low cost, such as Support content or product specifications. Some require high quality and creativity like Copywriting and Transcreation and you may even choose not to use TM’s for some of those. Some will be specific to parts of your Organisation while other will be global material. You will need to ensure a consistent Corporate identity across all these, in all languages.
  2. What are the fields: automotive, medical, IT require linguists with different backgrounds and specialisation. Make sure you know all the areas of expertise to cover during Translation and Review. For some you might to add Subject Matter Expert (SME) review to the more common step of Linguistic Review. Review changes will need to be implemented, communicated to Translators, fed into the TM’s, but the process will need to let SME’s take part in the process without having to learn CAT Tools.
  3. From a technical point of view you will also need to work with the content creators to determine the type of files you will receive from them and those they expect to receive back.
  4. Start a war on spread sheets as soon as possible. You probably won’t win it but the more you root out, the better. Teach your customers to understand how parsing rules protect their code by exposing only Localisable content during translation. Promote Localisation awareness during Development and Content creation. Document best practices such as avoiding hard-coded strings, providing enough space in the UI to accommodate the fact that some translations will be up to 30% longer than source text, at least if that is English.
  5. Your aim should be:
    • to receive files that can go straight to Translation with minimum pre-processing
    • to deliver files that your customers can drop into their build or repository for immediate use.
  6. No one should be doing any copy-paste engineering, manual renaming or file conversion.

Designing your Workflows

This can start with a pen and paper, a white board or whatever helps you think quicker, but it should end with a flowchart or set of flowcharts describing the process you’re setting up.

  1. Collaborate with your internal customers. You need to agree a signoff process, and avoid multiple source updates during or after the Translation process.
  2. Enumerate all the stages required and determine the following:
    • How many workflows do you need to describe all scenarios? Try to find the right balance: fewer workflows ensures efficiency, but too few workflows will lead participants to implement their own sub-processes to achieve their goals and you will lose control and visibility.
    • What stages do you need? The most common are:
      1. Pre-processing
      2. Translation
      3. Linguistic Review
      4. Post-Processing
      5. Visual QA
  3. Who are the owners of each step? Are they internal or external (i.e. colleagues or service providers)? How will you monitor progress and status? How will you pay?
  4. Is there a feedback loop and approval attached to certain steps? Will they prevent the workflow from advancing if certain criteria are not adhered to? Is there a limit to the number of iteration for certain loops?
  5. What automation can be put in place to remove human errors, bottle necks and “middle men” handling transactions.

Choosing your Vendors

Once you’ve determined which of your workflow steps need to be outsourced, you will need to select your providers. Linguistic vendors will likely be your most important choice.

Translation

In-house translators are a luxury rarely afforded. When choosing Translation vendors, first decide between Freelancers and Language Service Providers (LSP). Managing a pool of Translators is a job in itself, so most will hire the services of an LSP which will also be able to provide relief in terms of Project Management, Technology changes, Staff fluctuations depending on activity or holiday periods etc.. Having more than one LSP can be good strategic choice: it gives you more flexibility with scheduling and pricing. You can specialise your vendors according to content, region or strength. A certain amount of overlap is necessary for you to be able to compare their performance and benefit from a bit of healthy competition.

Linguistic review

Whichever setup you have for Translation, you will need linguistic review in order to ensure the integrity of the message is kept in the target languages. You will also need to ensure consistency between Translators or Agencies, check Terminology, maintain TM’s and Style guides.

Marketing and Local Sales Offices often get involved with that. However using internal staff removes them from their core tasks, unless you are lucky to have dedicated Reviewers. More than likely in-country colleagues will find it difficult to keep up with the volume and fluctuations of the Review work and ultimately will prove an unreliable resource. The solution is to hire the services of professional Reviewers. Many LSPs provide such services.  Some ask their competing providers to review each other, but that often results in counter productive arguments. A third-party dedicated review vendor will be the best to enforce consistency, accurately measure quality, maintain linguistic assets, and even manage translator queries on your behalf.

Selecting Technology

Translation Memory technology is a must. Which one you go for may be determined or influenced by existing internal processes, particularly if there are linguistic assets (TM’s and Glossaries) in proprietary formats. Your vendors may also have a preferred technology or even propose to use their own. If you go down that road, make sure you own the linguistic assets. The file format is another choice that needs to be made carefully from the start. Open source formats may save you from being locked into one technology. However technology vendors often develop better functionalities for their proprietary formats. It can be a trade-off between productivity and compatibility.

The good news is that conversion between formats is almost always possible. This means migration between technologies is possible, but avoid including conversion as a routine part of the process. Even if it’s automated, having to routinely output TM in several formats for example, will introduce inefficiencies and increased user support requirements.

Translation Management Systems have become so common, some think they are on the way out. You will at the very least, need a Portal to support file transactions, and share your linguistic assets with all the participants in your supply chain. Emails, preferably automatic notifications, should be used to support the transactions, but they should be avoided when it comes to file swapping. FTP is a common option, easy to set up, learn and cheap to run, but it can soon turn into a mess and gives you zero Project Management visibility. In order to achieve efficient status monitoring, resource pooling and any type of automation, you should consider a Translation Management System.

Whether you go for the big guns like WorldServer or SDL TMS, or for something more agile like XTRF TMS, you will reduce the amount of bottle necks in your process: handoffs will go straight from one participant to the next. The Project Managers will still have visibility, but no one will have to wait on them to pass on the handoff before they get started. TM’s will be updated in real-time and new content will become re-usable immediately.

A few things to look out for in your selection:

  1. Less click = shorter kickoff time. Setting up Projects in a TMS is an investment. It is always going to be longer then dumping files on an FTP and emailing people to go get them if you look at an isolated Project. As soon as you start looking at a stream of Projects TMS makes complete sense. Still, a TMS’s worst enemy is how many clicks it needs to get going.
  2. Scalability: you need the ability to start small and deploy further, without worrying about licenses or bandwidth.
  3. Workflow designer: demand a visual interface, easy to customise which can be edited without having to hire the services of the technology provider. Don’t settle for anything that will leave out at the mercy of the landlord.
  4. Hosting: weigh your options carefully here again. In-house is good if you have the infrastructure and IT staff. But letting the Technology provider host the product may a more reliable option. This is their business after all, maybe you don’t need to reinvent the wheel on that one.
  5. User support: the cost and responsiveness of the Support service is essential. No matter how skillful you and your team are, once you deploy a TMS to dozens of individual linguists there will be a non-negligeable demand for training and support. Make sure this is provided for before it happens.

Once you’ve made all these decisions, you will be in good shape to start building and efficient Localisation process. Last but not least, don’t forget to decide whether to spell Localisation with an “s” or a “z”, and then stick to it! 🙂

 

Related articles:

Crowdsourcing in Localisation: Next Step or Major Faux Pas?
Globalization – The importance of thinking globally
SDL Trados 2007: Quick Guide for the Complete Beginner
Which comes first, Globalization or Internationalization?
Who’s responsible for Localization in your organization?

 

Posted in Beginner's Guide | Tagged: , , , , , , , , , , , , , , , | 3 Comments »

SDL Trados Studio 2011 Preview: Can It Convince Trados 2007 Faithfuls?

Posted by Nick Peris on September 20, 2011

SDL have been drumming up interest for SDL Trados Studio 2011 through the summer. Eventhough the successor to SDL Trados Studio 2009 is announced to release at the end of September, I must admit that I have been slower to turn my attention to it than I was with Studio 2009.

This is in part due to my current occupation which brings me to spend more time using Translation Management Systems than CAT tools. But it is also because SDL Trados Studio 2009 was such an exciting breakthrough: the idea of fully integrating SDLX, Trados and Synergy was a major shift. The technology behind the new Studio file formats (.sdlxliff bilingual files, .sdltm Translation Memories, and .sdltb Term database) was also quite promising. Lastly, the productivity improvements were many thanks to the entirely new xml-based TM engine, which allows multiple TMs look-ups, AutoPropagation™, AutoSuggest™, QuickPlace™, Real-Time Preview etc.

Reading through those posts about SDL Trados Studio 2009 reminds me how attractive it seemed. But there was also a distinct possibility that this substantial innovation would not necessarily cause a mass migration of Trados 2007 users. Budgets were tight due to the worldwide recession. The prospect of migrating entire Localisation production chains seemed like an unnecessary overhead. Users would have to be re-trained, Enterprise and LSP proprietary automation redesigned in order to work with those new file formats. Above all, SDL Trados 2007 was delivering perfectly acceptable services.

Sure enough, two years later, empirical evidence suggests Trados 2007 is alive and well. It is apparent in my daily interaction with Localisation professional around the World. All Trados users are aware of Studio by now, but I’d venture to say all of them still have Trados 2007 installed, and that it probably even remains their SDL tool of choice. Assuming the hits on Localization, Localisation have any statistical value, it is a telling sign that SDL Trados 2007: Quick Guide for the Complete Beginner continues to be the most frequently visited post in these pages, 2.5 years after being posted. But then perhaps that’s my own fault, for not making a beginner’s guide to Studio 2009…

So let’s now turn to the future and look at SDL Trados Studio 2011’s prospects. New comers to the CAT tools market will inevitably consider Trados as one of their options; which new features it offers does not matter much. As for existing Studio 2009 users, I doubt any amount of innovation can make them upgrade if they haven’t already a budget or subscription plan which allows for systematic upgrades. The real measure of the impact of Studio 2011 will be whether it can convince the remaining Trados 2007 users.

What does SDL Trados Studio 2011 bring to the table to meet the needs of this demographic?

Some New Features

All the great advances made with Studio 2009 are of course still available, although some of them have matured. The main highlights in terms of novelty are the return of Perfect Match and the focus on productivity during review cycles.

Perfect match 2.0

Perfect Match makes a return to Trados: it existed in Trados 2007 but was absent in Studio until now. It now co-exists with Context Match, and together with Terminology and Sub-Segment leveraging make up the concept of Total Leveraging.

The differences between Perfect and Context Matches are:

  • Perfect Match can run on a batch of files (right-click a bilingual file to pre-translate and select Batch Tasks > Perfect Match) and is good for Project rather than document updates.
  • SDLXLIFF, TTX and ITD are all supported.
  • Context Match runs on successive versions of the same file, file names have to match.
  • They are marked as PM and CM respectively in the resulting bilingual files. Both segment types are locked.

Track changes

Studio 2011 uses a change tracking technology which is fully compatible with Microsoft Word. Thanks to the SDL XLIFF Converter, an SDL Open Exchange add-on now included in Studio, changes and comments made in Trados can be viewed, accepted etc. in Microsoft Word and vice versa.

This makes it easy to collaborate with users who do not have Studio during the review process. Whether they are linguists using other CAT tools or Subject Matter Experts not familiar with any CAT tool, they will all be able to input their feedback using Word.

The versions of Word officially supported are 2007 and 2010; 2003 should work but this is unconfirmed for now. Track Changes can be turned on or off for different parts of the process such as Translation, Review or Signoff under Options > Tools.

Display FiltersSDL Trados Studio 2011 Display Filters

In Trados Studio, segments can be filtered to show only those relevant to the current task. The filters in this list are another way Studio 2011 helps productivity during review, with new options such as Segments with Comments or Segments with Track Changes. These filters can also be applied during export using the SDL XLIFF Converter.

Improved Spell Checkers

Trados Studio 2011 brings the Microsoft Spell Checker back. Hunspell is still available but users can now configure which checker to use for each language. This is to resolve issues present in the Studio 2009 Spell Checkers which were not fully accurate for certain languages, notably Scandinavian ones.

SDL Trados Studio 2011 QA Checker 3QA Checker 3.0

QA Checker 3’s claim to fame is the interactive dialog box which makes reviewing and implementing reported issues a much clearer process. It is reportedly also a first step in longer term plans of adding grammar checks.

Enhanced File Filters

Studio 2011 includes new filters for:

  • OpenOffice, Libre Office, StarOffice and IBM Lotus Symphony.
  • INX and Java properties.
  • improved FrameMaker MIF support.
  • bilingual Word files which can now be edited directly.

Other novelties to discover in Trados Studio 2011 include pseudo-translation, for testing parsing rules and settings before the launch of new Project Types. Character, rather than just wordcount is now also available.

An Evolving Image

Lighter Ownership Experience

First impressions tend to last, and the installation and activation process are a big part of how a new application is experienced by users. In Studio 2011 the installation is made simpler. One single installer enables compatibility with Trados 2007 file formats (.ttx, .itd, TM upgrades and alignment tasks). With TTXit!, freely available on SDL Open Exchange, users should no longer need a copy of Trados 2007 in addition to Studio.

Because the user interface and technology in Studio 2011 are so similar to Studio 2009, no big learning curve is required. Any time and effort invested in learning to use Studio will just give users a head start in being proficient at the new version.

SDL Trados Studio 2011 MultiTerm WidgetStarting a project itself is a simpler process, with only 3 files needed (source, bilingual and TM), and no associated folder structure in the background.

The standalone License Manager has been replaced. Activation is now fully integrated into Studio, and borrowing licenses are supported.

Finally, the SDL Multiterm Widget is being pushed into the limelight. This taskbar tool lets you browse Terminology from external applications like Microsoft Excel, Powerpoint etc. at the touch of a button. It also provides a handy shortcut to searches in Google or Wikipedia and is now included in Trados Studio.

Expanding the Trados Community

Technology webinars have been an SDL strength for a long time now. Call it free education or a carbon-conscious alternative to business trips, they are an efficient way for any technology vendor to showcase their goods.

There are other ways SDL share information about Trados like the Studio 2011 Series on the SDL Blog, or the SDL Trados Youtube channel. SDL are certainly not the only language technology provider to use new media but I think it’s fair to mention their consistent effort to meet their user community and ensure information is widely available.

SDL OpenExchange is also used to promote this spirit of community with Developers (look out for prize competitions!) and has produced a number commercial as well as free Apps which efficiently respond to very specific needs.

The connectivity with SDL’s Enterprise applications is also kept up to date. Studio 2011 can connect to WorldServer or TMS Translation Memories for Concordance just like it would with local TMs. An Express Edition of Studio 2011 will be released for users who need Studio only for WorldServer projects.

Posted in News, SDL Trados Studio 2011 | Tagged: , , , , , , , , , , , , , , , , , , , , , , | 9 Comments »

Offline TM Update Process for SDL TMS

Posted by Nick Peris on July 19, 2011

SDLX Dashboard

While this is as efficient as one could expect, there are cases where “manual” updates of the online TMs must be performed. A Terminology change may have to be implemented globally into legacy TUs. A linguist may be asked to perform an offline clean-up of an overgrown or aging TM, and the resulting file may have to be imported back into the online TM. Audits may be conducted on live content and also require manual edits of the online TMs, etc.

In most cases, these edits will need to be performed by accessing the remote TMs using SDLX rather than SDL TMS 2007.  This is because SDL TMS doesn’t let linguists directly edit TMs, as I previously explained. The present post describes the step-by-step process to update TMs hosted on an SDL TMS 2007 server, using SDLX 2007. It can be used by linguists such as Translators, Reviewers or Language Leads or by Engineers depending on who in the process is in charge of implementing manual edits such as global updates or imports.Adding the SDLX Server Object

Prerequisites:

  1. SDLX 2007 Professional: no access to remote TM Servers is possible for SDLX Light or Freelancer users.
  2. TM Editing rights have to be granted to the users by the SDL TMS Administrator

Process:Opening a TM located on an SDLX Server

  1. Got to Start – All Programs – SDL International – SDL Trados 2007 – SDLX and start SDLX
  2. In the SDLX Dashboard, click Maintain
  3. In SDL Maintain, go to Tools – Options – Advanced – Object Management, click SDLX server and OK. Click OK again to close the SDL Maintain options dialog (this step is only required the first time you connect to an SDLX server)
  4. In SDL Maintain, click TM – Open – SDLX ServerSelecting the SDLX Server
  5. In the Select SDLX Server dialog, click Add and enter your SDL TMS server connection details
  6. Once the connection is established, open the Translation Memories drop down menu and select the TM to edit. Click OK twice to validate your choice and close all dialog boxes.
  7. Once the TM is loaded:
    1. Perform Text Searches by pressing F7 and edit as required (this is faster than using Find)
    2. Or import into the TM by clicking TM- Import
    3. Save and Close the TM when completedAdd SLDX server

Posted in SDL TMS, Translation Management Systems | Tagged: , , , , , , , , , , | 3 Comments »

SDL Studio Online 2011: the New Face of TMS

Posted by Nick Peris on July 12, 2011

Hot on the heels of SDL TMS 2011 which was recently reviewed here, SDL Technologies released SDL Studio Online 2011.

In a nutshell, SDL Studio Online 2011 is an optional add-on exclusive to SDL TMS 2011. It is distributed as part of the SDL TMS  2011 Service Pack 1 and upgrades the SDL TMS Translator Interface. While SDL TMS 2011 introduced a new Carbon Theme, its Translator Interface was in fact still a slim version of SDLX 2007. SDL Studio Online 2011 replaces it with an SDL Trados Studio inspired successor.

From a linguist’s point a view this is an important leap forward. Together with the performance improvements promised with the original release of SDL TMS 2011, this could significantly increase the proportion of linguists working online. Translators, and sometimes even Reviewers, have tended to choose to download SDL TMS packages rather than working online. This somewhat defeats the purpose of having an online translation environment with real-time TM and Terminology updates capabilities. Yet it has remained a popular choice mainly for two reasons:

  1. Combined server and user-side performance issues: this should already improve with an upgrade to the original SDL TMS 2011.
  2. Translation Interface too basic compared to desktop CAT tools: this is what this optional SDL TMS 2011 SP1 add-on proposes to address

SDL Studio Online in SDL TMS 2011Studio Online provides a number of tangible improvements over the standard Translation Interface:

  • Improved performance for linguists working online with the more modern interface featuring:
    • Segment-level lookup
    • Concordance lookup
    • Flexible tag display and editing
    • Find and replace functionality across the entire task
    • Integral spell checking with Inline and batch spell checking, multilingual dictionary and correction and user-specific dictionaries.
  • Reduced need for training and support for the growing number of linguists unfamiliar with SDLX
  • Reduced file management overhead thanks to a more attractive online environment
  • Increased proportion of linguists connecting to online linguistic assets rather than working from periodic downloads
  • Licensing and upgrade management owned by the client
  • Support for Microsoft input method editors (IMEs) for typing East Asian characters on non-matching language versions of Office.

Requirements and setup information

Studio Online is a Microsoft Silverlight plug-in. It requires Microsoft Silverlight 4 and the latest available Service Pack for Windows. Its upgrade or installation process is guided in Studio Online.

Other requirements are Microsoft Internet Explorer 6.0, 7.0, or 8.0 and Mozilla Firefox 3.5 or 3.6.

Enabling SDL Studio OnlineStudio Online Licenses have to be purchased in addition to the SDL TMS licenses for a specific number of concurrent users. Once installed and licensed, each user may choose to configure SDL Studio Online as their default editor in SDL TMS by going to Home – My Details – User Preferences and checking the option Use SDL Studio Online (see screenshot). This remains optional, so not all users have to make this choice. Task download for offline work also remains available in any case.

One limitation to note is that SDL Studio Online 2011 does not integrate with the QA Models. If you have made use of the online rating function added since SDL TMS 2007 SP4, you will have to make a choice between that and the new online interface. My inclination would be towards allowing the Translators to use Studio Online to make it more acceptable for them to stay online throughout the translation process. On another hand, Reviewers could continue using the old interface so that they use the QA Model, and so that you save on Studio Online licenses.

Posted in News, SDL TMS, Translation Management Systems | Tagged: , , , , , , , , , , , , , , , , , , | 5 Comments »

Kilgray TM repository: a New Home for Translation Memories

Posted by Nick Peris on July 5, 2011

Kilgray TM repositoryAs Kilgray Technologies made memoQ 5.0 Release Candidate available for download right on queue last week, there is another piece of Kilgray news I’d like to share with you.

The lesser-known but aptly named TM repository was launched recently by the makers of memoQ and offers an interesting and fresh approach to Translation Memory server products. This application apparently pre-dates memoQ but wasn’t launched commercially until this year. Since then, Kilgray have been gathering early adopters feedback, which they are planning to include in a version 2 sometime next year.

TM repository is made-up of 3 components: the database, the business logic and the web-based interface. It is built on SQL technology and comes in 2 editions depending on the number of users required.TM repository Importing Sessions

The idea behind any TM server product is to provide a central location where all users in a supply chain can access the same and latest version of Translation Memories. Different Localisation Managers have different TM Strategies which are often dependent on the CAT tools or even the version of the CAT Tools in use by the Assets owner and their LSPs. Important choices have to be made in terms of Maintenance, most of which have to do with how best to archive TMs for re-use. Working from project-specific TMs only gives smaller leveraging power and little version control ability.  Yet it is sometimes the chosen path, simply because it seems more manageable. On another hand building and maintaining Master TMs containing all segments ever translated, or even chunks of them organised by Product lines, Business Units etc. requires a sustained management effort. For instance, when there are terminology updates a linguist should implement global changes by batch editing Translation Units. They may spend time fixing old Translation Units (TUs) which will never be used again. It may also be difficult to find linguists with the skills to directly edit the TMs for all languages. More often than not, Master TMs which are not integrated with a Translation Management System will contain errors, deprecated terms, duplicate TUs with alternative translations etc. and require clean-up. The Project TMs-only route will always underperform in terms of ability to re-use existing translations and ensure consistency, but the Assets owner are still left to evaluate for themselves which option is the best for them.TM repository Maintenance Sessions

TM repository is a solution to a lot of these common problems:

    • It enables the Assets owner to create a single Online TM Database containing all TUs, for all projects, and all language pairs.
    • The flexible descriptive fields (metadata) allow the TU’s to be tagged precisely.TM repository Queries 2
    • This metadata can then be used in Queries for smart filtering during Maintenance or Export
    • TMX Imports let users add to the database from virtually any system
    • TMX Exports  permit the extraction of Project TMs, which can be reimported after use and update
    • Exports can be customised for the CAT tool in use through customisable Mapping. Query results (i.e. Project TMs) will contain metadata compatible with the target translation tool.TM repository Queries
    • Refined Maintenance is enabled through features such as Search and Replace of text or metadata, or the use of deprecation settings by which older TUs can be hidden from search results.

Posted in Kilgray, News, TM repository | Tagged: , , , , , , , , , , , , , , , , | 1 Comment »

SDL TMS 2011: Inner Peace

Posted by Nick Peris on June 28, 2011

The pace of release of Enterprise Technology such as Workflow and Translation Management Systems is usually slower than that of end-user applications such as CAT tools.

The reasons for this are easy to understand:

First, the priority for Enterprise Applications is stability, not cutting-edge User Experience. Users, and especially customers, require proven and sturdy environments capable of consistently handling massive traffic. This cannot be compromised in favour of the latest UI bells and whistles, not even the newest linguistic asset formats or features.

Secondly, the licensing and pricing model for these applications is such that customers have to monitor their ROI more carefully. Purchasing decisions would not be influenced with yearly or even bi-yearly releases of brand new product lines. The expectation is that these Applications provide a permanent Solution which can be used for several years to come. For that reason, Support contracts tend to include free Patches and even Service Pack upgrades which take care of the more pressing updates.

Last but not least, the efforts required in deploying these server-based technologies are again prohibitive of frequent upgrades. There are  infrastructure implications like matching SQL server versions or multiple server roll-out workload. The technology also needs to co-exist with a number of desktop applications in use in the supply chain.

From reading the SDL TMS 2011 Release Notes, I think the differences with its predecessor, SDL TMS 2007, are very much inline with these requirements. It seems to deliver relevant compatibility updates as well as promising improvements in usability and performance. If the announced increased reliability delivers, then I think one of the key to its success over SDL TMS 2007, will be whether it succeeds at making linguists more willing to work online, or whether they will continue to prefer to use it for File Transfer only and perform the actual linguistic work in their desktop CAT tools.   SDL TMS 2011 Carbon Theme

Compatibiliy

This is the first major release of SDL TMS since the acquisition of Idiom by SDL back in 2008. Together with the recent release of SDL Worlserver 2011, this confirms that in the short to medium term at least, these two Workflow systems will continue to coexist.

The SDL TMS offering features updated compatibility both in terms of CAT tools and infrastructure:

  • CAT: SDL Trados Studio 2009 SP3, SDL MultiTerm 2009 SP3/SP4, SDL Passolo 2011 (incl. word counts accurracy, new dedicated Workflows) and SDL Trisoft
  • Infrastructure: LDAP enhancements, Windows Server 2008 and Microsoft SQL Server 2008 support

SDL TMS can be upgraded to version 2011, though only from SDL TMS 2007 SP4 or later. SP4 would have to be installed first, before upgrading from any older version.

One piece of good news is that no data migration is required when upgrading, and all Post-SP4 hotfixes are included in the Upgrader. Microsoft .NET Framework 4 is recommended. 

Usability and Performance

The User Interface has reportedly been made more responsive in several areas: Translation Interface, Job Authorisation, Configuration edits and more. The UI has been updated with a new colour theme, but apart from that the navigation appears to be unchanged.SDL TMS 2011 Go to Dialog We will investigate in an upcoming article how this may be changed by the addition of SDL Studio Online. SDL Studio Online is an optional web-based version of SDL Trados Studio 2011, exclusive to SDL TMS 2011 SP1.

The Search feature has also been improved with increased speed for the main Search (results are now limited to 2,000 matches) and a new “Go to” feature lets users directly open specific Jobs or Tasks if they know the ID.

SDL Trados Studio can now access SDL TMS directly for TM Concordance and updates. This is achieved through an SDL Open Exchange plug-in. Once installed, users simply need to login using the SDL TMS Server Name, Username and Password, much like previously in SDLX’s SDL Maintain.

Unfortunately, Tageditor’s TTX files can’t be downloaded from SDL TMS 2011.  SDL recommend downloading Packages, which contain the ITD files for translation in either SDLX 2007 or SDL Trados Studio 2009. Eventhough SDLX is considered a part of SDL Trados 2007, this makes using Tageditor and Workbench more difficult and more-or-less means support for Trados 2007 in SDL TMS has been dropped.

Terminology imports have been enabled through a new functionality similar to the TM import added with SDL TMS 2007 SP4. This works using SDL MultiTerm .xml import files and a matching database definition. Passolo Terminology (sequences and TB updates) is also supported.

Here are a few other bug fixes and new features which caught my attention:

  • Users can reset their own passwords, which should improve the quality of life of many Workflow managers
  • Issues with the second and further pages of the Translation Interface have been fixed (comments, segment history and MultiTerm matches now work)
  • TM attributes can be edited from the Edit TM page
  • Ampersand(&) and quotes(‘) in ITD names are allowed
  • Job-level Project TM availability can be displayed in the Inbox
  • PowerPoint SmartArt is supported

Reliability

SDL claim that over 200 reported issues have been resolved, including a number reported by users through ideas.sdl.com.

Improvements in file format support and exception handling should limit the number of failed Jobs and Tasks.

Importantly, progress seems to have been made with Translation Memories exports. A new incremental method, saves having to use server resources to repeatedly perform full exports. This Incremental TM Export option, which is unchecked by default after installation, functions as follows:

  • Only segments added or modified since the last export are exported.
  • They are added to the latest corresponding TMX export file.
  • All TMX export files can be downloaded at once.
  • Note: segments deleted from the TM are not removed from the export. A full export (by temporarilly unchecking the Incremental Export box) is required to reflect any deletion

Posted in News, SDL TMS, Translation Management Systems | Tagged: , , , , , , , , , , , , , , , , , , , , , , , , , , | 2 Comments »

memoQ 5.0: Mr. Q Brings Change Management to the Localisation Continuum

Posted by Nick Peris on June 21, 2011

 
Mr.Q presents: memoQ 5.0!Kilgray Translation Technologies introduced memoQ 5.0 to the World last week by means of a twin event. Gábor Ugray, Head of Development, hosted a webinar from the Kilgray HQ in Budapest for the online enthusiasts, while István Lengyel, COO, demo’ed it live from the Localization World 2011 conference in Barcelona.

MemoQ 5.0 will be available as a public Release Candidate on June, 30 2011 and should reach Final Release within a few weeks of that.

The Release Candidate version can be installed side by side with memoQ 4.5 and various upgrade paths will be available to current memoQ users.

Following the strong focus on Project Management in memoQ 4, the philosophy behind memoQ 5.0 is Change Management. Changes in source files are better managed through X-translate, while segment changes are tracked through a sophisticated versioning system. Illustrated examples of this and other new features are detailed below.

memoQ 5.0 Version Tracking

X-translate

The implementation of Major/Minor version control is powerful because of the simplicity with which it responds to a real need. A Translator is working on a file, receives an update to the source file, thanks to memoQ 5.0’s Major versioning feature, he or she can immediately generate an updated version of their bilingual file and continue translating.

There is no need to leverage, which would require a more labor intensive process of pre-translating again from Translation Memories. One can simply go straight from a partially translated copy of version 1.0 to a partically translated copy of version2.0.

The screencaps below show how to xTranslate a single file from the previous Major version of the file, then how the  xTranslated segments are marked and finally how to save a snapshot of the resulting file.

xTranslate1xTranslate2xTranslate3

It is also possible to export a 2-column file for comparison of 2 Major versions:

Export 2 columns to HTMLSide by side compare

Change Tracking

Change tracking enables segment level access to previous versions. The following images show how to enable custom track changes from the Translation menu, how the changes are highlighted in a document, and a further 2 options for translators and reviewers to see changes made to a file since they last edited it.

Track ChangesTrack Changes Against BaseTrack Changes (Reviewers)Track Changes (Translators)

Terminology in memoQ 5.0

Terminology extraction

MemoQ 5.0 will allow a substantial amount of Terminology work without requiring the use of a dedicated application such as qTerm. Users will be able to extract candidate terms from a Project:

Extracting Candidate TermsTerm Extraction Progress

Stop Words

The use of Stop Words list will ensure easy noise reduction by preventing words such as “and”, “the”, or any other short listed by the user, from appearing as Candidate Terms:

Creating and Editing Stop Word Lists

Reviewing Candidate Terms

Candidate Terms can then be reviewed in context and possibly against an existing Termbase:

 Term Extraction ResultMerging Candidate TermsAccepted TermsDropped Terms

Lexicon

The Lexicon option will let you work with a Terms list without having to go through the full process of creating a Termbase. It is meant as an easy-to-use, immediately rewarding tool to manage Terminology within a Project. This should encourage Linguists to run quick Term extractions before starting a job, especially in cases where a Termbase is not available as part of the Handoff, in order to efficiently get a general overview of the Terms contained in a set of source files.

MemoQ 5.0’s Terminology feature does not support the TBX format, however Kilgray’s fully-fledged terminology tool qTerm, does.

memoQ 5.0 and nested file formats

Another very effective idea implemented in memoQ 5.0 is the support for file formats containing code belonging to other file formats. An obvious application is the case where the handoff is a spread sheet containing strings copied from an xml or a software file. But there are other common cases such as XML files containing HTML code.

The requirement here is to parse files twice so that all codes are recognised as such and so that the linguist can concentrate on translating with full confidence that all tagging is managed by the CAT tool. Here are 2 examples:

Cascading Filters

      1. Cascading Filters for a spread sheet contain HTML: 
        HTML code in XLS - ExcelHTML code in XLS - memoQ 5.0Reimport As to Apply Second FilterAdding a Cascading HTML FilterDocument Import SettingsSaving Filter Configuration for Re UseFully Parsed File
      2. Cascading Filters with Regex Tagger for a spread sheet containing UI strings: Run Regex Tagger to re-Parse XLS FileRegular Expression PatternsAdding Patterns to Configuration

Source Content connectors

Finally, memoQ 5.0 will also in time be able to connect to repositories where content is dynamically added. It is designed with CMS integration in mind, however the CMS connectors will only be released later this summer, like the web-based editor webTranslate.

Posted in Kilgray, memoQ, News | Tagged: , , , , , , , , , , , , , , , , , , , , , , | 4 Comments »

Alchemy Catalyst 9.0: A Practical and Visual Guide

Posted by Nick Peris on November 15, 2010

I recently had the welcome surprise of finding an invite to a Catalyst webinar in my Inbox. It was with great anticipation and a touch of nostalgia for my Localisation Engineering days, that I clicked on the link and joined the meeting to discover what Alchemy had been up to.

I soon realised that a practical user’s guide would be the best way to cover this on Localization, Localisation. The Alchemy Software Development website already lists What’s New in this release so rather than analysing the differences between Catalyst 8, for which we did a complete Launch coverage and Catalyst 9, I’ve put together a step by step tour based on the demo.

This article can be used by Localisation Engineers and Translators alike to preview the Catalyst 9 interface using the 30 or so screen shots included (see after the slideshow for full screen versions), and also to read through some recommended processes and tips, adding to my past article on the Leverage and Update Experts.

This slideshow requires JavaScript.

Creating a Project

The User Interface remains the flexible and now very familiar .net window, with its various docked panels and tabs. It’s also a stable interface which will cause little or no navigation headache to even the most novice user.

The first operation when getting started with Catalyst is to create a Project file, or TTK file. This is easily done by using the File – New menu and following the basic steps.

You will notice in the screen shots that the example used includes varied sample files such as compiled help (.chm) not requiring any source or project files, and wpf executable.Locked strings

Preparing a Project

After the creation of the TTK, source files can be inserted either using the Insert Menu item or a context menu in the Navigator tab. Folder structures can also easily be used.

Once the files have been inserted into the TTK, it is time to prepare it for leveraging.Translator Tool Bar Context Menu and Keyword Lock This operation of consists mostly of locking non-translatable strings and sub strings. It can be tedious on a brand new Project but the work done can be completely leveraged to the various language TTKs as well as any future versions of the project.

The lock keywords functionality has been improved in Catalyst 9: the txt file which the project’s keywords list is now automatically generated in the background as soon as the user locks a keyword.Catalyst 9 UI Batch Keywords Locking

Once a keywords list has been created, it can in turn be used to automatically lock the listed keywords in the remainder of the project.

Another thing to note is that Maximum String Length can now be set on a batch of strings at once.

Leveraging previously translated content

Apart from Leveraging from the TTKs of previous projects, Catalyst supports leveraging from a variety of Translation Memory formats:Keywords List

  • Translation Industry Open Standard (*.tmx)
  • SDL Trados 2007 (*.tmw)
  • Wordfast Pro (*.txml)
  • Tab-delimited (*.txt)
  • Alchemy Translation Memory (*.tm)
  • Alchemy Catalyst (*.ttk)
  • Alchemy Publisher (*.ppf)

Alchemy Translation Memory is a new proprietary format used to create Master TMs from completed TTK projects. This format allows to store Catalyst-specific context information such as the context (Dialog box ID, Menu Item etc.), which can later improve the quality of leveraging by providing Perfect match. In Catalyst terms, a Perfect Match is a 100% match located in the same Dialog, Menu etc).TM Compatibility List

Alchemy Publisher, Wordfast Pro, Trados 2007 or the nonproprietary TMS are also present provide compatibility with other TM format Catalyst might have to coexist with.

Noticeably, Trados Studio 2009 TMs (.sdltm) still do not appear to be supported.

Batch processing

The process recommended by Alchemy is to create an English to English Master TTK and then to automate its duplication and pre translation for each target language in the Project.

This is an area where Catalyst 9.0 does seem to bring a good bit of novelty:Create Job Expert

  • With Catalyst 7, engineers had to manually duplicate TTKs.
  • Catalyst 8 was a bit more helpful and created Project folders for target languages and project resources.
  • In Catalyst 9.0 however, the Job file and Scheduler take care of a lot of the repetitive tasks associated with preparing a new Project.

The Create Job Expert lets you use the Master TTK as a template to create project folder structure and corresponding target language TTKs.

Meanwhile, such tasks can also be added to the Scheduler. This new queuing system allows the user to start working on the next project while it processes queued tasks in the background.Create Job Expert Batch Leverage

Automation

The Command line automation has been improved since Catalyst 8 to include Analysis. The complete Catalyst localisation process can now be automated.

Catalyst 9.0 Developer Edition also includes the Comm API which lets advanced users script TTK operations all the way down to string level, and output automation reports  in txt or xml format.

Ensuring Quality and Consistency

In addition to Translation Memories, Catalyst 9 also supports several Glossary formats:

  • Text files, used in Catalyst since the beginning (.txt)
  • Terminology Exchange Open Standard (.tbx)
  • Translation Memory Exchange can also be used for Terminology (*.tmx)
  • SDL MultiTerm and MultiTerm ServerCatalyst 9 inline Validation

Validation still takes two forms: the Expert can be run to perform global check, and inline validation can also be switched as a non-intrusive real-time quality control. If a potential error is found, a flag will be raised through the bottom pane, but Translators will not be interrupted. They can simply go back to the issue by clicking on the notification once they are ready to attend to it.

The Thumbnail view seems to be a great tool for engineers regressing bug. It gives a preview of all dialogs in a TTK and lets you click the one which matches for example the screen shot in a QA report and brings you automatically to the location of this dialog in the ttk file.Catalyst 9 Thumbnails

Translating in Catalyst

The Concordance search and Translator toolbar do not appear to have been changed. Both were introduced with Catalyst 8 where there was strong focus on improving the user experience from the Translator’s point of view, and they seem to have delivered.

The new Re-cycle button is a result of the same ambition. New translations can be propagated to entire project by using the current project as an inline TM in the background. Layouts are not recycled but fuzzies are supported.

Clean up Expert

Finally the Clean up Expert has also receive some improvements. Like for all Experts, it is recommended to close the Project file before running it, and then select the file(s) to process from the Expert’s General tab.

Clean up now creates a postproject.tm Translation Memory and generate supplied assemblies for .net.

Conclusion

in my opinion, this new generation of Catalyst still offers a great solution for visual localisation. Although the differences with Catalyst 8 may not may not make a bullet proof case for immediate upgrade, the 25% discount currently on offer does represent decent value.

Posted in Beginner's Guide, Catalyst, News, Software Localisation | Tagged: , , , , , , , , , , , , , , , , , | 1 Comment »

SDL TMS 2007 SP4: Some Comments from SDL

Posted by Nick Peris on June 2, 2010

SDL TMS

Here are some interesting comments from Paul Harrap, Product Manager for TMS at SDL, in reply to my article on SDL TMS 2007 SP4. I’ve also included my own response afterwards.

“(…) I’m very pleased to see we’re getting some coverage in the blogosphere. I’d like to thank you for taking the time to write us up and your positive feedback on the product generally and our new SP4 specifically.

I accept that there’s still some work to do in the product with relation to TM maintenance. As your article accurately reflects, the contents of TMs are updated in TMS in very specific places in the workflow- typically after one or two cycles of review – and what content goes into which TM can be carefully controlled. This is very much by design. We see Translation Memory as the crown-jewel of the linguistic assets of the enterprise customer and so contents are tightly regulated by TMS.

However, we have to acknowledge that bad content can creep into TMs over time – there might be an error in review, or some customers might not review translations quite as thoroughly as others. The changes we made in SP4 to allow the import of files directly into TMs is a response to this requirement. The enterprise can now add/replace contents of a TM directly, without reference to a specific translation job or workflow, as an administrator-level function. This can allow people to quickly and painlessly correct known-bad TUs.

We’re considering including the ability to search through, browse and directly edit the TUs in the TMS browser environment in a future release. While I accept that this is a lacking feature, I wouldn’t concur that we should be putting such power in the hands of the vendor or the freelancer. Seeing the TM as a hugely valuable asset for the enterprise, I expect this is the sort of feature and capability that most enterprises would want to keep in-house.

On the integration with SDL MultiTerm, I very much see a distinction at the moment where TMS is a consumer of Terminology and MultiTerm is the owner of it. Over time we will see much tighter integrations between the SDL products, so the lines between TMS and MultiTerm will very much start to blur, and we have plans to introduce workflow capabilities for term lifecycle management.

On the issue of uploads and downloads and working offline, I think a lot of people would very much agree with you. The single largest corporate user of SDL TMS is… SDL! We have dozens of translation offices around the globe, all of whom deal with the upload and download of files to and from TMS servers based in our hosting centre in London on a daily basis. What tends to drive people offline is the featureset available in the desktop tools. SDL Trados Studio, and its predecessors SDL Trados TagEditor and SDLX, are very powerful productivity tools for the translator. Replicating these features in an online translation environment is a monumental task and it’s  something we are investigating.”

First of all, I would like to thank Paul for this input. Since the ramp up of Trados Studio over a year ago, SDL have made a sustained effort to listen to their user base. The TMS section proves here that they are keeping with this policy.

On the topic of TM Maintenance, which is very close to my heart, I think the business model Paul is presenting is either slightly outdated or, more likely, is missing on a part of their customer base.

From my experience, the outsourcing model has developed so much during this recession that at least in some cases, big enterprises (i.e. the TMS customers) no longer employ Translation Memory management experts. These positions are filled by technicians employed by the LSPs.

Another point is that while TMS customers use Review routinely, they also cannot afford to review all the content they output. Most of the big players have either implemented or are looking into models which allow them to reduce their review cost for languages where the quality is considered stable. This means that TMs may be updated in TMS with contents which hasn’t been reviewed, and consequently that linguists must regularly inspect the TMs and fix any inconsistencies in legacy TUs to prevent reoccurring errors.

I maintain that this task must be assigned to a linguist, and the best placed to do so is a senior Reviewer. Not all of these are in-house, by far.

Posted in SDL TMS, Translation Management Systems | Tagged: , , , , , , , , , , , , , , , , | Leave a Comment »

SDL TMS 2007 Service Pack 4: Love and Hate

Posted by Nick Peris on June 1, 2010

SDL TMS 2007 - Localisation workflow

I always find it challenging to get a fair idea of what Enterprise tools can do before making a purchase decision. There is so much involved in setting them up that even if a trial version is available, the efforts required to perform meaningful testing are prohibitive.

Many such applications do not come ready out-of-the-box and require extensive customisation before they can be tailored to fit a specific business model.

This is why many purchase decisions are executive decisions, based on ROI reports and presentations showing what the software does. A demo might be setup for you on a dedicated server by the sales person, and you’ll be left thinking “hum…surely it’s not that simple”. This is also why 10 times out of 10, these pieces of software come with a Support package which lets you install regular and much needed updates and bug fixes.

It doesn’t have to be this way!

If you have the opportunity, go knock on a few door and try to find a company nearby which uses the software in a production environment. Contact them, ask to visit, get an independent demo. From my experience (not based on TMS that time) most people will be more than happy to tell you how much effort it took to setup, how many features still don’t work, but also how much their productivity has really increased and perhaps even how many of their employees have done a thesis on the subject! Bottom line: get real-life advice!

SDL TMS, or Translation Management System, is one such behemoth application. Trying to find independent information about TMS on the web is a challenge. In fact, even finding official information can prove frustrating. As for Special Interest Groups… those I found were for customers-only. It seems it’s buy first, we’ll talk later.

So what’s the big deal exactly? Well I’ve been working with TMS 2007 for about a year now and I have a few things to report: some good, some not so good.

What it does well

Let’s start with positive thoughts.

TMS is a workflow tool, designed to connect a customer directly to it localisation vendors and all their armies of sub vendors. It handles big volumes and short turnarounds really well, and is reasonably good at supporting your Translation Memory and Terminology Management needs. It also offers the reporting facilities necessary for all members of your localisation ecosystem to invoice each other, and you.

TMS automates part of the role of the middle men, and is ideal for localisation consumers with a constant stream of translation, especially if they come in the shape of numerous small projects.

Multiple alternative workflows can be set up, depending on vendor selection, TMs to leverage against, TMs to update, need for Linguistic Review etc. Once the correct workflow is selected at the job creation stage, you can be sure it will go through all the steps required. There is little or no human error possible, at least not in scheduling and assigning tasks to the right participant.

TM updates are handled automatically, literally seconds after the last human input in the workflow.

Where it lacks

So are all the vendors really gathering orderly around the assembly line and localising thereafter like a happy family?

Not exactly. There are a few snags.

My main grief is around TM Maintenance or the lack of it. Because TMS automatically updates the Translation Memories at whatever stage of your workflow you told it to, manual editing of the TMs has been neglected. A user can perform a Concordance search, but it is impossible to edit the Translation Units found. One cannot use TMS to fix inherited inconsistencies or any error found in legacy TUs.

This makes implementing Global changes a very untidy task: one needs to connect to the TM Server (hosted by SDL in most cases) using SDLX 2007 Professional. This, to me is total non-sense and here is why:

  1. increasingly, the business model in Localisation is outsourcing.
  2. once localisation is outsourced to agencies, these subcontract Single Language Vendors, who themselves might only be sub-contracting to freelancers.
  3. less and less Localisation consumers employ in-house linguists.
  4. their remaining in-country staff is Sales and Marketing, and has much more pressing matters to attend than editing TMs.

Now which version are these freelancers more likely to have? SDLX 2007 Professional (€2,995) or SDLX 2007 Freelance (€760)? I think you probably guessed it. SDL’s licensing model prevents linguists from maintaining TMs in TMS and seemingly forces corporations which bought TMS to support their outsourcing setup, to fix TMs in-house!

There are some workarounds to this, but for a piece of software of this caliber, I think this is a pretty shocking limitation.

The integration with MultiTerm has similar issues: only some of the functionality are available through TMS, the rest including editing Term entries has to be done using MultiTerm Online or Desktop.

Performance issues also tend to drive a lot of linguists offline! Depending on their setup, a lot of them find it more efficient to download jobs, translate offline in SDLX and upload the finished work back into TMS. While there is technically no difference in the end result, this is a disappointing interruption of the workflow.

Service Pack 4: An End to the Suffering?

Squeezing under the gate at the last second, like Bruce Willis in a classic movie, TMS 2007 Service Pack 4 sneaks in before the long-awaited SDL TMS 2010 and comes to the rescue.

With TMS 2010 now possibly slipping into 2011, it is a welcomed addition particularly due to the improvements it brings. Here are the most significant end-user facing features:

Browser support: IE 8 support added (IE 6 removed in future)

TM import: ITD, zipped ITDs, MDB (SDLX TMs). This is a partial solution to the lack of TM Maintenance feature I’ve talked about in this article.

Continued lack of support for TMX is attributed to the fact that this open-source format has too many proprietary specifications.

Reporting formats added: CSV, Excel 2007, PDF, RTF, Word 2007.

Branding and Fonts are customisable (by Professional Services).

TMS 2010 is expected to have end-user customisable reports.

Segment level QA Model for Reviewer grading

QA Models

This all-new feature in SP4 is crucial if your workflow includes Linguistic Review. All changes made by the Reviewers are now recorded, and the Reviewers can tag them using customisable Error Rating and Categories.

  1. Error Ratings and Categories: support for LISA model, SAE J2450, TMS classic out-of-the-box.
  2. User-specific models can be created. Number of points deducted can also be specified in the QA Model.
  3. Records can be retained at segment (for feedback to translators) or project level
  4. Scoring methods: absolute or percentage
  5. To apply a QA Model: add it to a Configuration (i.e workflow), and it will be available to Reviewers working on jobs passed through this config.
  6. Reviewer usage: click Star at segment level to open the QA model window and enter Category and Rating.Pass/Fail status does not prevent reviewer from submitting or rejecting a job.

Posted in News, SDL TMS, Translation Management Systems | Tagged: , , , , , , , , , , , , , , , , , , , , , , , , , | 5 Comments »

memoQ 4: Interview with István Lengyel

Posted by Nick Peris on December 22, 2009

I have been trying to diversify the topics we cover on LocLoc; and especially the tools we talk about. It started recently with a QA tool and now continues with a CAT tool. I already know from the survey I’ve had on this page, that a lot of you are familiar with Kilgray’s memoQ. This, is a preview of what to expect from the forthcoming memoQ4, from the mouth of Kilgray’s COO, István Lengyel.

[Nick Peris] Hi István, could you introduce Kilgray and your role within the company?

[István Lengyel] Hi Nick! Thanks for inviting me to do this interview. Kilgray Translation Technologies is an independent company dedicated to the development of clean and innovative tools for translation, but so far we are by far the best known for our memoQ translation environment. Though we are based in Hungary and all the founders are Hungarians, we became quite an international team in the last two years, opening up in Germany, Poland and now in the US. It’s really great to work in this team, as we have people coming from all sorts of companies such as Idiom, Passolo, SDL Trados, etc., and every addition to the team opens up new perspectives and shows new approaches – the company culture builds on respect and cooperation.

I am one of the architects of memoQ and also the chief operating officer at Kilgray, though in reality I’m mostly managing our sales and marketing team and our international expansion.

[Nick] Could you give a general overview of what memoQ is for readers who are not familiar with it?

[István] memoQ is an integrated translation environment that has a couple of focal points. First, it is easy to use, easy to learn. Second, we translate a lot in it and manage memoQ’s localization in memoQ itself, so we developed an eye for details – there are lots of smaller features that really make life easier. Third, from the very beginning we were concentrating on collaboration, and even the first version included an internet-enabled TM/TB server. Fourth, we don’t believe that we should lock in any of our customers – the entire system supports interoperability between tools to the maximum extent, meaning that you can process files prepared by virtually any major translation tool, and you can also prepare files for processing in other tools. There’s also a full set of documented APIs available for integration with other tools. Fifth, leverage, which means that we are trying to make the most of your resources. There were a couple of things where memoQ pioneered: we were the first to introduce real-time previews that change as you type, we were the first to introduce communication such as knowledge bases and instant messaging and offline synchronization into a translation memory server, we were the first to introduce the translation memory-based segmentation where pre-translation emulates the way your translators join and split segments, and we were the first to introduce the automated concordancing. But quite frankly, we are just as happy to take over things that work from other tools as we are to introduce new stuff.

[Nick] I know you are preparing to release a new version; could you give us a release date for memoQ 4?

[István] A few days ago we named January 31, 2010 for the release date, but I was reminded that it’s a weekend. So the first week of February. (Well, who cares about weekends? :))

[Nick] What are the main changes from memoQ 3.5 and main reasons to upgrade?

[István] There are so many changes that I can hardly list them! memoQ 4 is the first memoQ version that really focuses on project management. We like to build bottom-up and believe that an organization will only have a good experience deploying a tool if the translators like it, and we spent the last five years making the translators happy. So let’s start with the revolutionary feature: post-translation statistics. Imagine a situation where several people are working on the same set of similar documents, using a server-based translation memory. There can be a lot of fuzzy matches coming from the other translator’s translated entries, but so far there was no way in any tool to enumerate these matches, because the person who starts working later gets more matches than the person who is the first to start. memoQ 4.0’s post-translation statistics will solve this Gordian knot, and give you the actual fuzzy match analysis for every translator after the project. This way finally there is a business model for server-based translation.

Other than this, the biggest change is that we have upgraded the concept of translation memory servers to the concept of resource servers. So far you could share translation memories, term bases and documents between translators, and you could set up projects for them centrally. In the new version, you can share every other resource such as auto-translatables (for people used to Trados lingo: customizable placeables), non-translatables, segmentation rules, QA settings, keyboard shortcut settings, ignore lists for the spell checker and so on – 12 of them, all together. What’s more, sharing this happens in the background so you can start the publication of a big TM on the server and go on managing other projects in the meantime. These resources can all be exported into an XML-based format so clever project managers can prepare them also automatically.

memoQ 4 also brings finally the concept of multilingual projects. You can create handoff packages and receive delivery packages, or you can simply publish a project on the server. Those who receive the handoff package can in turn create new handoff packages (handy for a multi-tier enterprise-MLV-SLV-translator setup), and through delivery the files and reports are updated automatically. The handoff packages are just zipped containers of open-source format data – XLIFF for documents, TMX for TMs and CSV for terminology. You can process the packages in any tool, so the users are not locked in.

Compared to these improvements, the brand new text editor, the completely revamped user interface and the streamlined quality assurance seem small. Even the previous version of memoQ got quite a lot of credits for its good support of bidirectional and CCJK languages, memoQ 4 takes this further and also introduces support for Indic languages. We are introducing a very advanced multi-tier undo/redo logic, real-time spell checking and other minor improvements. The quality assurance checks have also been dramatically improved and also the interface for fixing warnings has been fine-tuned.

And I failed to mention so many things! memoQ 4 is the single biggest upgrade memoQ ever received.

[Nick] For non-memoQ users, could you give us the main reasons to switch to memoQ 4?

[István] Because other people do and they are happy about it! 🙂 Just like every company, we make mistakes at times but there has not been any single case that anybody asked for a refund. Seriously, I think the main reasons to switch to memoQ are collaboration, interoperability and support. memoQ is a truly collaborative application, it is one of the few tools that enable simultaneous translation and proofreading on the same document, complete configuration of projects for your translators, or using several translation memories or term bases that can be local, remote — they can even be on different servers — or offline synchronized. The server is fast even on a HSDPA connection and it’s also very affordable – no wonder we have over 150 servers out there.

The other important aspect is interoperability. Our main market is language service providers, and an LSP can never say that they use only a single tool, period, otherwise they lose business and what’s more, they can also lose translators. With memoQ you can process documents and packages created by other tools, and you can prepare packages in industry-standard formats for other tools too. Therefore you don’t find yourself in a situation that you bought the tool because you liked it and then you have to fight with everyone around you to make it accepted.

And the third most important aspect is support. I think Kilgray’s support is just great – fast, focused and friendly.

[Nick] What is the pricing structure for memoQ 4?
What are the different Editions of memoQ 4?

[István] memoQ 4 comes in three client editions: translator standard, translator pro and project manager.

memoQ translator standard is for those translators who never work in teams. It does not enable access to servers and does not enable export of files into XLIFF or bilingual DOC, only memoQ’s proprietary MBD format. It also lacks the ContexTM (101%) matching which takes the context also into account, and comes without support. But the price tag is attractive: 99 euros a year.

The memoQ translator pro is the edition for professional translators and very small translation companies who don’t want to invest into a server solution. It costs 620 euros.

The memoQ project management edition comes with multilingual project management and reporting functionality and we charge around a thousand euros for that.

When it comes to server technology, we sell our solution with mobile (ELM or floating) licenses, meaning that companies can give away and take back licenses to translators over the internet. The initial package contains five mobile licenses, and we sell additional bundles of five licenses at very competitive prices. When it comes to servers, we prefer not to sell without a trial period of 30 days – we want everybody to use the tool, not just buy it for the drawer.

[Nick] How did you take into consideration user feedback during the development of memoQ 4?

[István] Oh I could name the people who contributed with their user feedback here! I think it’s worth mentioning how we work. Basically there are four people who decide on what gets into the next release, and every release has a theme. These themes are contained in our 5-year roadmap and we regularly come together for things that we call “walk in the woods”‘ – creative sessions outside the office where we discuss the main ideas and concepts. We personally talk a lot with users and try to learn the rationale behind their feature requests. These talks shape the main themes/features a lot. On top of that, we have a system to archive all the threads on feature requests, and we go through these regularly. I could give you a rather precise list of features for the next three versions!

So basically the user feedback is taken into consideration on two levels: when we realize that a business problem is hard to solve with memoQ, we incorporate the solution into the high-level concepts. The other level is the feature level where for example users request amendments to file filters or suggest small usability improvements. If these are justified, these can go straight into the feature overview.

[Nick] How is Terminology Management undertaken in memoQ 4? What are the Termbase formats supported?

[István] Terminology management is one of the most controversial components in memoQ! So far we only support CSV and – surprise-surprise – TMX as import formats and can also export into Multiterm XML. Why TMX? Just think about software localization and then the help and you’ll understand. With memoQ we decided that this is a translation tool and not a terminology application, and therefore we gave a finite set of attributes but something that is pretty comprehensive: you can have synonyms, definitions, notes, grammatical information, contexts, project, domain, subject, client information, and a few other fields. You can also have images in the term base, and forbidden term variants can also be flagged. From the workflow point of view, memoQ has had a term base moderation feature since v2.0 in 2006, which means that terminologists may need to approve all terms suggested by translators before they become final. Terminology matching is really exciting: you can use wildcards to indicate the end of the invariable part of every word in a term, i.e. for a language like Spanish you can enter cinturón* de seguridad and that will also find cinturónes de seguridad. For translators of Slavic languages this is really crucial (fuzzy matching does not always work for terms). I can list quite a few pros for memoQ’s terminology management but I must say that it’s a very practical approach. However, we understand that corporate terminology management is not a subset of translation, and terminologists may need some more freedom.

Expect that freedom in a third-party tool based on the memoQ engine soon.

[Nick] Is there anything specific to memoQ in the way Translation Memories are created and maintained?

[István] Translation memories are by default context-enabled in memoQ, and memoQ supports two kinds of contexts: the segment before and after and context bound to structural information. This latter means that if you have for example the software strings in an XML or Excel file, with an attribute indicating where the text appears, you will get a 101% match if the attribute is the same to the attribute where you originally entered this translation – this way you can shuffle the translatable strings and still keep the context information. If you speak the Idiom lingo, this is very similar to ICE and SPICE matching.

As for maintenance, there are a couple of things that are quite unique. First, a 100% or 101% match for us is only a match that is identical both in content and formatting to the original. But we have a special bracket, 95-99% that contains segments where numbers, formatting, whitespaces, punctuation marks can be different. Any change in the text results in something lower than that. You can join and split segments wherever you want, and when you get an update to the document, the TM-driven segmentation will automatically join and split the segments according to your previous translation, as it looks into the translation memory for better matches through joining and splitting. During pre-translation, cases where you get multiple 100% matches (because you translated the segment differently in two contexts, and this third context is unknown so far) are flagged and they are very easy to locate. All these features fall under the umbrella term we use for design: “reproducibility”. I think it’s also worth mentioning that memoQ has a built-in TM editor and can work with as many TMs at a time as you wish. Oh, and yes, a minor nuance, just to make things elegant and please those who are really tech-savvy: our support for TMX also covers attributes, so if you import a TMX file coming from another tool that has attributes, even if the TMX attributes there cannot be displayed in memoQ, you can expect that the TMX export from memoQ will preserve and contain them – so memoQ does not swallow the information that it cannot process.

[Nick] Is there any new feature in memoQ 4 you are particularly fond or proud off? Maybe some anecdote about features which took you a lot of efforts to achieve and which you are now very happy to bring to memoQ 4 users?

[István] Well, I’m a person who prefers the big picture to the small details, and for me the biggest achievement – and a big praise goes to Gábor Ugray, our head of development who designed these features – is that the tool did not get more complicated for translators according to the feedback of those users whom we showed the system. We always pay a lot of attention to the user interface, but when we started conceptualizing memoQ 4 about two years ago, keeping its simplicity seemed like a daunting task. The visual marker of the entire resource management and multilingual project management feature is now just two drop-down lists: the server selector and the language selector. And I am of course proud of the fact that the resource concept makes the entire system future-proof – no matter what sort of a linguistic resource comes into existence in the next years, we’ve got a place for it, and savvy users are also welcome to write third-party resource managers.

[Nick] We are seeing a merging trend where tools are less specific to either software or documentation. This is partly due to the content types evolution, and partly to an effort by tool developers to become more all encompassing. How does memoQ fit into this? How is your support for software localisation? Also xml and xliff?

[István] I saw this very much in 2005 when we started off but I don’t see it that much anymore. About a year ago or so we implemented visual localization support for RESX files and quite a few users are using it, but we have no plans to implement visual localization for other formats such as RC or binary files. On the other hand there are quite a few considerations in memoQ that make it a very good tool for localizing Help content. I already mentioned the TMX import into the term base and the support for context based on another column in the Excel file or an attribute in the XML file, I’d like to mention the automated concordancing feature that was inspired by one of our translation jobs – in our earlier lives as translators – where TM management (another issue I could talk about for hours) was virtually non-existent. I don’t want to name the end-client and the LSP we got this from (they are both very reputable and well-known in localization), but basically to translate the help of version 8 of a well-known application we only got a TM that contained version 2 to 7 of the same application. No terminology, no localized software strings for version 8, nothing. We spent hours to find out what screen caption has been translated before and what expressions did we have to coin, because – as it is with software – quite a few of them were 8-10 words long, and of course developers make changes to these every now and then, changing one or two words maximum, adding a few words to the end, etc. The automated concordance automates this manual process: it automatically gives you the longest multiword expressions that appear at least a given number of times in the translation memory. It does not give you the translation in most cases, but if you select it, it opens the concordance window with the right expressions. And yes, the concordance can look for a series of words. So basically we don’t want to take away business from the excellent software localization tools, but we definitely want to be the best technology for translating help and manuals.

[Nick] Do memoQ and Kilgray offer workflow technology allowing supplier and clients in the localisation chain to work together online?

[István] Our workflow is a linguistic one, and not a highly structured one. We coined two terms. For us, horizontal workflow means when people work together on the same task. Vertical workflow is the traditional workflow, passing along the files between different people doing different jobs. memoQ is excellent in helping people work together on the same task and has a lot of workflow tools such as moderated term bases, simultaneous translation and proofreading, different forms of review, communication and knowledge bases, etc. From the point of view of traditional workflows, we only cover translation and review – items that happen within the tool. There’s no way to integrate things like source text review, DTP or settlements into memoQ. However, the extensive set of APIs enable integration with workflow tools, and at this point I have to mention that both Beetext Flow and Plunet Business Manager do a great job when it comes to deep integration. They can both take care of the entire process, and generate and maintain the projects automatically in memoQ. One of the things we are putting a lot of emphasis on nowadays is client review. I think memoQ is one of the best tools for this, but there is still a lot of room for improvement.

[Nick] Could you say a few words about the memoQ support network? How can new users avail of the experience of other users and if necessary receive support from Kilgray directly?

[István] Here are a couple of interesting resources: http://rc.kilgray.com – the Resource Center that contains training videos, guides, filter configurations for XML-based file formats, but also interesting articles on general topics such as TM management, technology purchase pitfalls, etc. for people and companies not using memoQ.

The memoQ Yahoo! Group (http://tech.groups.yahoo.com/group/MemoQ/) offers the expertise of other users but we also contribute often, and hey, you have the best experts of the competition also there and they often contribute too.

There is a memoQ wikibook too, and the forums on proz.com and other sites can also be interesting.

If direct support is required, it’s primarily through our support email address – please don’t publish the address directly on your website, we don’t want more spam there, but it’s at kilgray.com.

[Nick] Is it too early to ask you about roadmap? What are you plans for memoQ?

[István] It’s not too early at all, but I’m afraid I can’t tell much about the big improvements at this point. One thing is for sure – after 4.0, we will relax a bit and iron out any rough edges that may have remained in this brand new tool. One of the things that many users asked for and will be there in 4.1 (or whatever the final version number will be) is the bilingual DOC table format for review with comments. But one thing is for sure, you can expect another major version with a huge new resource in 2010.

[Nick] This has been a very informative interview. I thank you for your time and detailed answers and look forward to reviewing memoQ4 in the new year!

Posted in Interviews, Kilgray, memoQ | Tagged: , , , , , , , , , , , , , , , , , , | 3 Comments »

QA Distiller 7: Sanity Checks on Steroids

Posted by Nick Peris on November 17, 2009

QA Distiller is a great quality control tool I came across when I was working on the Marketing project I already mentioned in an article about XML in Localisation.
Developed and distributed by Yagamata Europe, this tool has a lot to offer to client-side engineers, multilingual vendors and freelancers alike. In fact I was even using it to enforce proper and consistent use of Terminology in source marketing content, before sending for localisation.

With the impending release of version 7 at the end of this month, I thought it was the perfect opportunity to talk about it on Loc Loc. The purpose of QA Distiller is to batch process quality checks on bilingual files. Essentially, it performs similar tasks to the QA Checker in Trados‘s TagEditor, but with some major differences.

The benefits

Multiple file processing: QA Distiller allows you to run a highly customizable list of checks on batches of files. There is no need to open of each individual TTX file, or run the QA Checker successively on each one. Just select the files to process, the settings to apply and run the tool to output a comprehensive report for your follow-up. This is a great way to control and enforce consistency across entire handoffs or projects. Translation quality, Terminology consistency etc. are simultaneously audited across all the files selected.

Multi-lingual processing: better yet, this can also be done across all languages at once, which is particularly powerful for controlling Do Not Translate instructions have been adhered to, for example.

Interactive reporting: the report output is another great selling point. It rates and classifies errors and lets you update it as you review and fix or discard candidate errors. It can be exported to a variety of formats where source, target and error details are summarised and categorised. This is very helpful to communicate with vendors on queries, as well as measure the quality od deliveries. Finally, the report has hyperlinks not only to the file, but to the actual segment where the potential error was detected. This makes the implementation of fixes really quick and easy. No more peeling your eyes out to find typos or endless finger-cramping Ctr+F session. If there is an error, QA Distiller will get you right there!

Software stability: my experience (version 6 in Windows XP) has shown very solid performance and compatibility, and certainly far less crashes than SDL’s QA Checker.

Some rare shortcomings


One of the limitations I found in the current version was that the Translation Consistency check did not work when running QA Distiller across several languages. Instead of reading the language code of each file and filtering the comparison, it reported the fact that translations differed from one language to the next. Not particularly helpful.

Secondly, although the pricing structure offers good choice, the full version seems a bit steep at €1000, especially since it also requires Trados to function on TTX files.

Additionnal Technical Information

QA Distiller supports all languages, and a variety of file formats: TRADOStag documents (TTX), FrameMaker RTF (STF), Translation Memory eXchange (TMX).
Terminology can be checked against proprietary-format dictionnaries (DICT) or the industry-standard Term Base eXchange (TBX).

The upcoming version 7 introduces:

  • Tag and ID-aware terminology checks
  • New Wrench icon funcitonnalities: batch correction of multiple quotation mark and number formatting
  • Fine-grained ignore option for improved noise filtering
  • Tag and case-independent consistency check
  • Full support for Georgian, Malay (Rumi and Jawi), Serbian (Latin and Cyrillic)

The little green man also told me that there are plans to add support for the many different XLIFF flavours like SDL XLIFF, MemoQ XLIFF, WorldServer XLIFF by the first quarter of next year.

For more details, check the cool demo at http://www.qa-distiller.com/movie/‏

Posted in QA Distiller, Quality Management | Tagged: , , , , , , , , , , , , , , , , , , | 8 Comments »

SDL Trados Studio 2009: The Compatibility Questions

Posted by Nick Peris on July 21, 2009

SDL Trados Studio 2009: the Compatibility QuestionsOne by one, SDL continue to address obstacles to our upgrade decisions. Earlier today, one of their webinars tackled the critical topic of Compatibility in Translation Supply Chain. A recording will be available at www.sdl.com, but here is a quick summary for our convenience.

Alignment

This feature is not included in the recent release of SDL Trados Studio 2009. It is planned as an upcoming update, but until then SDL Trados WinAlign or SDL Align from the Trados 2007 Suite must be installed to perform alignment work.
Once the alignment performed in Trados 2007, export to Translator’s Workbench TXT, and import into a Workbench or SDL Maintain TM respectively.
That TM can then be upgraded to a Trados Studio 2009 TM (.sdltm). The export can also be imported directly into an sdltm but with implications regarding TM settings (see section below).

Translation Memories

Upgrading old TMs

Old TMs and bilingual files can be upgraded (File and Server-based supported).
TMX can be imported directly into sdltm, but if imported into tmw or mdb first, the TM settings can be imported into the sdltm.
3rd party TMX may not support this path fully, as they may contain settings specific to the CAT tool used to create them.
The Upgrade Translation Memories wizard in Trados Studio 2009 can batch process various TM formats for various language pairs simultaneously. A Custom option lets you rename the TM output files if required. Segmentation rules can also be imported. Translation Units can be filtered out by field (e.g. you can choose not to include TU’s tagged as “Not approved”).
You can also choose to output as many as TMs as you input, or merge same language pairs independently of their TM formats.

Publishing new TMs to Trados 2007 users

Export to TMX ensures vendors and colleagues who are using Trados 2007 can use your TMs. But this removes support for features like Context Match because information on previous segment, style…is only carried in sdltm.

Creating new TMs from old bilingual files

Importing ttx or itd directly into sdltm (batch) allows Context information to be added (this process however will not carry over any TM setting). It’s a case of either or.

Termbase compatibility

MultiTerm 2007 Termbase can be opened directly and automatically upgraded by MultiTerm Desktop 2009 (it’s one-time operation obviously).
Termbase from earlier versions of MultiTerm have to be converted into xml first, using SDL MultiTerm Convert, before they’re imported into a MultiTerm Desktop 2009 Termbase (.sdltb). Selecting
Catalog-Export allows to export using the Default Export Definition, to export to xml (which can be used by a user with MultiTermT2007)

Bilingual files

TTX and ITD can be opened directly. They will be converted to sdlxliff (e.g. filename.doc.ttx.sdlxliff)
At the end of project they can be saved back to TTX or ITD using
Save Target As. Use select
Original
TRADOStag Document to save as ttx (requires SDL Trados 2007 Suite).

Tag settings files

SDLX and Trados settings files can be upgraded. To do so go to
Tools-Options-File Types- select file type and click Copy to duplicate the default settings file in Trados Studio 2009. Browse to save location and move your new settings file up and down the list to set its priority against the original one. Import the legacy settings

into the new File type setting you created.

TMS and MTO

Compatibility with SDL TMS will be implemented, as soon as integration is available. The date is unconfirmed and SDL advise to stick to SDL Trados 2007 Suite TM, Bilingual and TermBase formats for the time being.

Compatibility with MultiTerm Online will only occur with the upcoming release of MultiTerm Server 2009.

Posted in Beginner's Guide, SDL Trados, SDL Trados Studio 2009 | Tagged: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , | 1 Comment »

Globalization – The importance of thinking globally

Posted by Patrick Wheeler on May 21, 2009

Crouching Tiger, Hidden Dragon…

In essence, Globalization (Internationalization in MS speak) is your Kung Fu. Bear with me, I have a point here, either that or this is a thinly veiled attempt on my part to get you to read further. 🙂

Globalization represents more than just an all-embracing term used simply to describe the sub-processes of Internationalization and Localization, it is in fact both an ethos and strategy that describes how your organization needs to position and prepare every facet part of its being.

Those familiar with Chinese martial arts or who have spent too much time watching Kung Fu movies will understand the fundamental difference between the Tiger fighting style and the Dragon fighting style. The Tiger style relies on sheer strength and the memorization of moves, whereas the dragon style is based on the principal of a deeper understanding of movement. It’s about anticipating more than simply acting upon and reacting to events.

Staying on the fortune cookie philosophy theme, if you adopt the Tiger approach to Globalization you may make all the right moves, correctly identify your target global markets, prepare and push forward with Internationalization of your product with vigour and determination, and skilfully and swiftly execute product localization, but even this is not sufficient if you want to ensure your business is ready to go global and prepared for the effects of going global.

You need to adopt the dragon Style. In addition to the above actions, you should seek a deeper understanding of the impact that these actions will have on your business and anticipate this reaction. After all, every action has an equal and opposite reaction. Once you have decided to go global with your software offerings, you will have to consider how this decision will subsequently impact all areas of your business such as Programme/Project Management, Development & QA, Sales & Marketing, Legal, Accounting, Distribution, Support, etc.

Thinking out loud – So who does what?

Product Management: will need to coordinate with all groups to ensure that localized releases are part of any global product roadmap and are approved by and communicated to all stakeholders.

All global product release schedules need to recognize that the Development and QA teams will have to work in “harmony” with Localization Engineering and QA, and therefore core Development and QA time and resources will have to be allocated to addressing I18n, Customizability and Localizability issues.

Failure to factor these tasks into any global project scope will mean that a simship will be impossible, Developers and QA alike will be frustrated by having to potentially allocate additional time to deal with unplanned for I18N defects, Localization will be stalled until defects effecting Localizability and Customizability are addressed, and regional sales channels will suffer from late availability of localized product.

Development & QA: As mentioned above, these core groups, usually charged with domestic software releases, will now need to work in-synch with their Localization counterparts; the frequency and format of handoffs to the Localization team need to be agreed, I18N exit criteria will need to be established  for design and development phases, pseudo-localized software builds will need to be created for I18n testing, code freeze dates will need to be agreed to allow for the extra volume of i18n defects that will be logged during I18n/L10n testing, the workflow and management of i18N defects through the core defect tracking system will need to be established, and core Development and QA resources will need to be allocated to resolving and regressing i18N, Localizability, and Customizability defects.

The Localization team will mainly be focussed on addressing L10n issues, so the majority of I18n and Localizability issues will need to be resolved by the core Development team.

Even prior to Internationalization, it is essential that those at senior levels within an organisation understand the impact of going global on their core Development and QA teams.

As highlighted in my first post, assuming that the creation of localized software releases is the sole responsibility of a single Localization team is imprudent and unrealistic. Globalization means a significant investment in core Development and QA time and resources and cannot happen in isolation of these groups or without their involvement.

Sales and Marketing: Sales and Marketing teams responsible for the target regions need to be made aware of strategic plans regarding localized releases. Often these groups will be the ones who identified the business case/requirement for a localized software release.

Regional Sales and Marketing teams will have an insight into the features that are important to their markets and any customer issues with in-market localized product that need addressing as a matter of priority for subsequent releases. They will also be able to advise on any region specific customization of software features that will be required. These customizations will need to be considered during design and development under the heading of “Customizability”. Furthermore, it is important for Programme Management to work closely with these teams when formulating the localised product roadmap, ensuring they are involved in any beta program review of the software and they have sign-off as part of the localized product review process. This may all seem fairly obvious and simply requires clear lines of communication, but I have often witnessed a certain disconnect between regional offices and global Programme Management.

The following excerpt from Beyond Borders – Web Globalization Strategies by John Yunker (2003) is a good example of how poor communication and planning within an organization can ensure a rather embarrassing false start on the journey to global domination;

“The marketing director of a professional society wanted to expand the subscriber base in other countries. The society already had many international members, but because none of the publications had been translated, members needed as least a moderate grasp of English to reap the benefits of joining. So the marketing director decided to translate the society’s membership form into Chinese, in the hopes that it would make joining the society much easier for Chinese speakers and increase membership.

Within a few weeks, the society received its first completed Chinese form by fax, the membership directory, unaware of what the marketing director had been up to, looked at this form, filled out in Chinese, and said, “What the hell am I supposed to do with this?” The membership director didn’t understand Chinese. No one of her staff understood Chinese. Even if someone on her staff did understand Chinese, their membership database didn’t accept Chinese characters.

So this person in China completed the membership form and subscribed to a couple of publications and the organization could do nothing about it. The professional society didn’t even know what publications were selected because the publication names were translated to Chinese – and they had no English template to compare it against. It may seem obvious that you shouldn’t create marketing materials in a language your company can’t support, yet companies that jump into global markets too fast frequently repeat this scenario.” (Yunker, 2003, p.82).

Branding and cultural customization are also important considerations that also require input from regional Sales and Marketing groups. Some may favour regional branding and cultural customization over global branding with a universally consistent user-experience. This allows regional Sales and Marketing the flexibility to better connect with their target audience. It is all too easy to alienate your customers if they get the impression that your organization’s software products, website, support etc were not developed with their region in mind. However, others would argue that allowing such distinct and unique branding combined with a high level of customization on a region-by-region basis, simply serves to dilute global brand power, resulting in a confusing and inconsistent user-experience. Additionally, by allowing diverse and inconsistent localized content per region, the global management of this content can be troublesome and costly.

The whole area of cultural customization is vast and there is a lot of information as well as misinformation offered on this topic, and it can be hard to discern urban legend from truth. On the theme of colour and cultural significance of colour in the global marketplace, one publication I read recently would lead you to believe that red cars are illegal in Brazil and Ecuador because of the perception that they cause more accidents. This is in fact absolute bunkum. So approach cultural customization with caution and seek the guidance of local contacts.

Legal: There are a variety of laws governing software being sold in different regions of the world, many of these laws pertain to language and support for the official languages in these regions; such as the Toubon law in France, GB18030 certification for China, and the charter of the French Language in Quebec (Bill 101).

For translation of End-User License Agreements (EULAs) and software warranties, your organization will require the services of legal translators and a review of the EULAs by your in-country operations centres/partners to ensure compliance with local legislation.

Legal regulation on the sale of software worldwide is unlikely to become any more lenient. To the contrary, with proposals such as the EU’s two year guarantee for software (games), which would allow users who are unhappy with “buggy” software to return their purchase, the situation will only become more complex. This is another reason why a well thought-out Globalization strategy combined with a strong focus on I18n is of paramount importance.

With poor I18n, your localized software will inevitably contain more functional and cosmetic defects than the source release, and that could be a real headache when faced with a future where customers are within their rights to simply ask for their money back on the basis of these defects and are not compelled to wait for a hotfix as may currently be the case under the terms of existing EULAs.

Accounting: Your accounting team must be ready to provide pricing in the local currencies of the regions your software is to be sold into. Accordingly, they will also need to be ready to accept payment in these currencies. Ensure you have a clear understanding of how royalties and revenues from localized software sales are distributed throughout your organization.

Distribution: You will of course need to consider your distribution channels, competition, and how you will physically deploy your localized software to your customers. For hosted solutions, automatic updates etc; existing data centres serving your domestic customers may not offer sufficient connectivity/speed to customers in other regions.

Support: Before you have localized software in-market, your organization will need to be ready to support these target markets. It is an all too common mistake to simply expect that this will somehow take care of itself and that existing support channels for domestic product will be sufficient. This is yet another way to disaffect the customers in new markets you’ve worked so hard beguile with your digital wares.

You need to consider the mechanisms for localized support; knowledge base, email, phone etc. What level of support will your in-country operations centres/partners can offer, if any? How are support issues with localized software escalated? Do your call centre representatives have the necessary language skills and knowledge of the localized software to handle calls/emails from all the regions you sell your software in? Do you have a Content Management System (CMS) behind your existing website/knowledge-base? Does the functionality of this CMS lend itself to the management of global content in multiple languages?

Once the knowledge-base route has been exhausted, there is a common preconception that it is a good idea to heard customers to email support, like cows being shoved into a cattle crush, as opposed to presenting them with the option of phone support. This is based on the logic that email support is far more cost-effective than phone support. Whilst it makes sense to encourage customers to avail of email support over phone support, I do not believe it is a good idea to completely eliminate phone support as an option.

Many organizations prefer to remove any reference to phone support from their site. For me, this represents a false economy, whilst you may be saving on call centre costs, you will probably be losing customers, and any chance of repeat business. This is particularly flawed strategy in new markets where you are fighting for market-share.

I have yet to experience an email support system where I have received a (useful) answer “within 24 hours” as promised. Besides, 24 hours may be a long wait depending on the nature of the issue. Even if there is a customer cost associated with phone support, it is better to offer this as an option as opposed to lose customers who may prefer to simply return your software (see “Legal” above) and align themselves with your competitors rather than wait for a delayed response from support.

What happened to Localization??

You may have noticed that I have made no mention of the Localization team/departments specific responsibilities in terms of Globalization. This is a deliberate omission. I will address aspects of Localization in various future posts (after all, the URL for this blog puts me under some pressure to do so!). For now, however, it is more beneficial to recognize that in the grand scheme of Globalization, Localization is actually one of the simplest components. Granted, as “Localization” experts, we are in fact required to be “Globalization” experts and provide guidance in relation to Globalization strategies, but if all other areas of your business are ready to go global, then Localization should be the least of your worries.

Once again, failure to take a holistic approach to Globalization will result in Localization being a tedious, costly, and protracted affair. Localized product quality will suffer and inevitably your organization’s performance in the target region will be poor. Additionally you will have filled the lives of your Localization team with a degree of despair! So for the sake of good Karma, get the fundamentals right and Localization will be a walk in the park.

The above are just some of the areas for consideration when formulating your Globalization strategy. One could certainly write a book on the topic and a number have been written on the topic. Globalization is the broadest and most subjective area when it comes to looking at G11n, I18N, and L10n and is therefore open to the most debate.

What color/colour is the sky in your world?

The Sapir–Whorf hypothesis (roughly) states that through the medium of language, different cultures attempt to define their reality and enforce a structure on the world as they view it. This results in certain perspectives that are unique to particular cultures; this is why Localization and Globalization extend beyond simple translation.

This probably also goes some way to explaining why a Chinese friend and work colleague of mine finds a particular Rice Krispies Squares TV commercial so amusing, whilst I simple perceive it to be mind numbingly boring. Or maybe I just don’t get it! Whatever the case may be, to be truly successful in a particular regional market, your organization will not alone have to speak the language of that region, but also understand the predominant cultural perspectives distinct to that region.

The important thing is to have a carefully considered Globalization strategy that would make Lex Luthor seem nonchalant in his scheming, and to execute the plan in a decisive and coherent manner throughout the organization and without procrastination. Understanding that Globalization is the responsibility of your entire organization and must permeate through every level is a good first step.

This is particularly important in the current economic climate. Whilst many organizations are running home for shelter and scaling back on their global operations, this presents opportunities for other organizations to get traction in emerging markets if their Globalization strategy is sound. It may be a long term investment, but if your competition is busy running for cover, these recessionary times could represent an opportunity to gain market share in valuable new markets. As Warren Buffett said, “Be fearful when others are greedy, and be greedy when others are fearful.” In other words, advance when your competition is retreating from global markets.

In conclusion, you could of course try the Tiger approach and see what happens, but as another icon of our times (Homer Simpson) once said, “Trying Is the First Step towards Failure”. 🙂 So instead I urge you to think like the Dragon and have a deeper appreciation of how Globalization will impact your own organization and how your organization as a whole will need to evolve to meet these challenges.

Posted in Globalization, Internationalization | Tagged: , , , , , , , , , , , , , , , | 1 Comment »

SDL @ Guinness: Trados Studio 2009 Q&A

Posted by Nick Peris on May 15, 2009

SDL Trados Studio 2009

The SDL Roadshow was in Dublin yesterday.

The “cream” of Ireland’s Localisation community was treated to a big day of product demos and slideshows at the home of the black stuff: the Guinness Storehouse.

As I made my way through Guinness town under a refreshing morning drizzle, I wondered for a minute how the pungent aromas of the early brewing activity would agree with the power breakfast I had had not so long ago.

This was soon to be forgotten however, thanks to a flying start to the proceedings provided by SDL’s Internal Training Manager, Tracey Byrne. Her performance was followed by a few other SDL presentations, as well as a case study on TMS by LSP partner VistaTEC. By the time we reached the Gravity Bar (it must have been 17:59) for some last minute networking opportunities, I think it’s safe to say we were all satisfied by a great event and a fine venue.

There was a lot of information provided throughout the course of the day and I will be releasing separate articles on SDL Passolo 2009 and SDL MultiTerm 2009 soon. What follows below is more directly related to SDL Trados Studio 2009, and what is new or adds to my Preview article. I’m presenting it in a Q&A structure which I hope will be practical to anyone looking for information on specific features, and an easy read for anyone wishing to go through it all. Sláinte!

What is the release date for SDL Trados Suite 2009?

The development cycle has reached Release Candidate stage and SDL are working towards an end of June release target.

Have the development team taken user feedback into account?

Yes, 80 ideas for Trados and 16 for MultiTerm are a reflection of user suggestions on ideas.sdltrados.com

Are TagEditor and Workbench gone?

Yes, Trados Studio combines aspects of SDLX and Trados into a fully integrated User Interface. Even MultiTerm, which still installs separately even though it is bundled with Trados Studio, now offers full functionality from within the Studio UI. SDLX, Workbench and TagEditor simply do not exist anymore.

What are the system requirements?

Here’s what SDL Marketing are saying on the subject of System Requirements:

“SDL Trados Studio supports Microsoft Windows XP and Windows Vista. As minimum requirements, we recommend a Pentium IV-based computer with 1 GB RAM and a screen resolution of 1280×1024. For optimum performance, we recommend 2 GB RAM and a more recent Pentium or compatible processor with a higher screen resolution.”

Please note that this is still subject to change until closer to the launch in June.

What is RevleX™?

It is a new XML-based TM engine. SDL Trados Studio 2009 uses new file formats for bilingual files (.sdlxliff), translation memories (.sdltm) and termbases (.sdltb). It brings together a slew of new features such as Context Matches, AutoPropagation, AutoSuggest™, Multiple TM support etc.

How does AutoSuggest work?

AutoSuggest is an inline predictive text-like feature which provides suggestion from TM, Termbase or dictionaries as you type. Suggestions appear in a context menu, with an icon clearly indicating whether they come from the TM or Termbase etc.The user can customize the maximum number of entries offered. Suggestions start appearing from the first letter typed and keep updating until you select one or finish typing the word.

Can you turn AutoSuggest off?

I’ve also heard this question about Alchemy Catalyst 8.0‘s ezType™. Perhaps from the same person?. The answer is Yes (in both cases), but developers have spent brain cells trying to make these features work in a non-intrusive yet efficient manner so you should probably give it a fair go!

Are Multi-lingual XML files supported?

Bilingual xliff will be supported but there seems to be a question mark on multi-lingual, and SDL said they’d follow-up with me once it’s clarified.

What are the improvements to format filters?

Main progress has been with PDF, XML, FrameMaker and inDesign.

How does the Upgrade TM functionality work?

Trados Studio will convert your old TM into the new format. In the current implementation this requires for the version used to create these TMs to also be present on the same machine. The alternative is to extract the TM on the machine that has the old version and import the content back into a Trados 2009 TM. I was also told that this may yet change and they may be able to include the components of the old version required for TM conversion in Studio 2009. Watch this space!

How is navigating big files in Trados Studio ‘s Editor improved compared to TagEditor or even Trados-aided Word?

The left panel in UI lists the headers and lets you click them to jump to a particular area in the document.

How does the Editor’s Real-Time preview work?

You need to manually generate the preview once. It uses a built-in stylesheet to simulate the end-result. This does not work on DTP file formats.

Can I lock segments in the Editor?

Yes. Context Matches (CM) are locked by default, but the PM can also manually lock other segments.

How is XLIFF supported?

Standard XLIFF are directly supported. The new default format for Trados bilingual files is .sdlxliff which is a proprietary format developed from XLIFF with additional functionality relating to RevleX™

How does QuickPlace work?

To apply formatting, highlight the word or group of words in the target segment, press CTRL + comma. Choose the required formatting from the inline dropdown list. If there is more than one to apply in a segment, QuickPlace will try to guess which is most likely required and offer it at the top of the list. Alternatively you can also hold CTRL, highlight the formatted text in the source segment, and then highlight the text to be formatted in the target segment. Similar applies to Placeables such as figures, measurements etc.

Is there Real-Time verification in the Editor?

Yes. If an error is detected, an icon will appear in the notification area between the source and target segments. The error message can be viewed in the tooltip of this icon or in a dedicated message panel. In case of False positive, simply remove the warning.

Does Trados Studio 2009 support TTX files?

Yes for editing, no for creating.

Is cross-files AutoPropagate available?

No, not in the first release. But there is a workaround: Merge all project files into one.  Cross-file repetitions are also taken into account when creating a package if the “Recompute” option is selected.

Does the Merge feature support all file types?

Yes files of different formats can be merged together. Once merged they can still be viewed and worked on relatively independently.

What is new with Term recognition?

The Editor allows direct access to full MultiTerm functionality. Terms can be cross-reference by ID so if a term is edited, any other term previously linking to it for definition remains linked.

What is the workflow in a scenario where not all participant to a project have upgraded to Trados Studio 2009?

If the Project Manager has upgraded the translators, reviewers etc will have to upgrade in order to use the TMs, to open the bilingual files or use the Termbase. The Project Manager will be able to work with Trados 2007 files (creates a .ttx.sdlxliff) but not create them.

The only alternative is to provide TMX translation memories and not to pre-translate the deliverables.

Can the PM upload project packages through FTP using the Project panel in Trados Studio?

No. Project packages can only be email through Outlook. This is however optional, and FTP can always be done manually once Trados Studio has created said packages.

Can you import customer details?

Yes but only from Outlook.

Can multiple TMs be used in a project?

Yes multiple TMs and Termbases are supported. A priority order between TMs can be set and there is also an option to “Always use the best match”.

What’s new with fuzzy matches?

The fuzzy band values and their number are now fully customizable.

What reference material can be included into a package?

Package can contain global TM settings, Termbases, AutoSuggest dictionaries etc.

Does Perfect Match still exist?

No, it is replaced with Context Match (CM) but may be added back in a later release.

What does Create Package do?

  • creates a folder structure
  • creates a package per target language if the option is selected
  • lets the user define tasks for individual packages
  • recomputes wordcount or analysis for cross-file repetitions.

Are files locked for updates while packages are out for translation?

No. It would be a good suggestion for ideas.sdltrados.com, to mirror a functionality in SDL Passolo 2009.

What is the LSP partner program?

52 Language Service Providers have entered various levels of partnership with SDL. The objective is to create value for translation buyers, help LSP’s become experts at translation technology, and promote training and support.

When will training for Trados Studio 2009 be available?

Training for SDL Passolo 2009 is available now. Courses (including upgrade courses) for Trados Studio 2009 will be available at launch. There will be a split between a Translators and a PM path. There will also be a separate SDL MultiTerm 2009 course.

When will certification exams for Trados Studio 2009 be available?

End of September 2009.

Which training and certification path will be on offer?

For Translators:

  • Getting Started
  • Intermediate
  • Advanced
  • MultiTerm

For Project Managers:

  • SDL Trados Studio 2009 for Project Managers
  • SDL MultiTerm 2009 for Project Managers

Posted in SDL Trados, SDL Trados Studio 2009 | Tagged: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , | 5 Comments »

Which comes first, Globalization or Internationalization?

Posted by Patrick Wheeler on April 8, 2009

In my previous blog entry, I covered the limitations of “Localization” as a generic label to describe what we in the software “Localization” industry continually strive to achieve under the headings of G11n, I18n and L10n, as well as the dangers of this branding in terms of how “Localization” can often be perceived as the sole responsibility of a single “Localization” group or department within an organization.

To add to the confusion, there are two separate and somewhat contradictory models used to describe the relationships between G11n, I18n, and L10n. Microsoft’s model and the model predominately used by the rest of the industry! J Naturally you will also encounter subtle variations to both these models within various organizations.

So before examining G11n, I18n, and L10n in more detail, it’s probably useful to familiarize yourself with the key differences and similarities between these two models.

Microsoft’s Internationalization Model

The graphic below (Fig. 1) represents Microsoft’s “Internationalization” Model.   

Microsoft's Internationalization Model

Microsoft's Internationalization Model

The main thing to be aware of, and where this model is at odds with the model used elsewhere in the industry, is in the terminology. In Microsoft’s model, the terms “Internationalization” and “Globalization” are substituted. “Internationalization” is seen as the overall, high-level process, and “Globalization” is a sub-process that deals with the development of a culture-independent/world-ready application. 

N.B. There is some inconsistency in terminology within Microsoft’s own documentation and content; “Globalization” and “Internationalization” are sometimes interchanged depending on the target audience, author, time of day, weather, etc.

The “Industry Standard” Globalization Model

On the other hand, the rest of the industry typically refers to “Globalization” when talking about the overall process, and “Internationalization” when describing the development of a culture-independent/world-ready application. See the more commonly accepted, “Industry standard” Globalization Model below (Fig. 2). 

The “Industry Standard” Globalization Model

The “Industry Standard” Globalization Model

The irony of this inconsistent terminology won’t be lost on anyone working in Localization. J

At first glance you may assume that Microsoft’s model (Fig.1) provides a more comprehensive description of the whole workflow, as there is more detail provided in the high-level model. This is not strictly the case. Whilst the more commonplace model used by the rest of the industry (Fig. 2) is typically only represented by three neat little Globalization, Internationalization, and Localization boxes, there will of course be more detail under each of these headings, but the level of detail/terminology will once again vary from organization to organization. For example, if we expand the model in Fig. 2 further, we would see something similar to the following workflow (Fig. 3) emerging:

Expanded "Globalization" Model

Expanded "Globalization" Model

In Fig. 3, I have placed “Localizability” and “Customizability” under “Internationalization”. In my opinion, these are just a few of the more significant component parts of Internationalization. If we were to expand the I18n process still further, one would see the addition of other major I18n considerations such as Unicode. 

Resistance is (sometimes) Futile

There is no right or wrong model to adopt or champion within your organization. Essentially both models describe the same overall process. However, it is useful to be aware of both models, especially if you have the misfortune of having to delve into Microsoft Documentation relating to Internationalization or the Globalization Namespace. Similarly, when talking to people from the Microsoft/.Net universe, I’ve found it can be easier to simply give up trying to stick to the more widely accepted G11n model and speak in Microsoft terms. Otherwise it can be rather like trying to convince the Borg there is an alternative to assimilation (I ‘m already sorry for that reference!) and you may find yourself viewed with the same skepticism as zoologist who just suggested polar bears and penguins could peacefully coexist. J Apologies to my ex-Microsoft colleagues, but you know it’s true! J

In my next few posts (and as previously promised!), l will endeavor to work-around the (at times) conflicting terminology and take a look at the commonality in what these process models are seeking to describe under the headings of Globalization, Internationalization, and Localization.

Posted in Globalization, Internationalization | Tagged: , , , , , , , , , , , | 1 Comment »

SDL Trados 2007: Translation Memory Strategies

Posted by Nick Peris on March 27, 2009

What is the best way to organise and maintain Translation Memories?

I currently maintain TMs using 2 features of Trados (the Attributes and Master/Project TM dichotomy) and Alchemy’s Trados component.

Master TMs

  • single and exhaustive repository for each field and language pair (e.g. EN-FR Medical).
  • used to analyse all new projects and generate Project TMs.
  • content of Project TMs are only added to it when full project cycle has ended (including review, QA etc.).
  • because of their exhaustive nature, Master TMs tend to grow rapidely and would not be practical for inclusion into a translation kits.
  • even when outsourcing all or most of the localisation process, these should always be held by the client as they are a valuable asset which they own, regardless of whether they are outsourcing TM Management.

Project TMs

  • specific to a project or project stage (i.e. successive handoffs of a same project often have their own Project TMs).
  • used to pre-translate the handoff (i.e. generate the TTX files to send to the vendor).
  • passed on to translation vendors for analysis and use during interactive translation.
  • used during post-translation engineering (bugs are fixed in Workbench + TagEditor + MultiTerm interactive translation environment by the localisation engineer).

Software TMs

  • single and exhaustive repository for each field and language pair, generated bi-yearly from Catalyst TTKs.
  • added to Master TM of their field and/or used as Concordance reference during translation of help, documentation, knowledge base articles etc.
  • also used as leverage source for software through Catalyst.

Use of attributes

  • every time a new project is analysed, custom attributes are added and set (e.g. Vendor=AAA, Project=XXX, Field=FFF).
  • can be used to filter searches and analyses.
  • also useful to track back on errors or arbitrate between duplications.

Posted in SDL Trados, SDL Trados 2007 | Tagged: , , , , , , , , , , , , , , , , , | 2 Comments »

Who’s responsible for Localization in your organization?

Posted by Patrick Wheeler on March 27, 2009

Who’s responsible for Localization in your organization?

Seems like a simple question with a simple answer, right? However, whether they are aware of it or not, most people use the term ”Localization” when they may well be referring to areas under the broader headings of Globalization, Internationalization, Localization & Translation (GILT).

There are historical reasons for this anomaly of course; once upon a time Localization was only considered an afterthought to product development and had no real place in the SDLC (Software Development Life Cycle). GILT is certainly a more accurate and all-encompassing acronym, but even as industry experts in “Localization” we do not typically embrace such broad terminology. Personally I find GILT a somewhat clumsy and uncomfortable acronym. After all, who in an organization would want to say they work in GILT, or are head of GILT! Even if we were to adopt this term within our organizations, I could foresee many blank stares when discussing GILT with those not familiar with what is traditionally known to them as “Localization”. So naturally we default to using “Localization” as an often all-encompassing term to avoid having to give every person we interact with a brief (and most probably unwelcome) history of what is better known as “Localization”.

The problem is, that by accepting our moniker as “Localization” we are also endorsing the view that Localization is still just an afterthought to development and is solely the responsibility of a single department within an organization. I still work as part of a Localization team, as Localization Engineering Manager. Some of you who work in the industry probably have a sign hanging over your little farm of desks that says, “Localization”.

In my experience, this tends to result in those in senior management, in charge of strategic decision making, and those in regional sales offices, believing that by having a Localization department; Localization is taken care of. It’s a black-box. It’s possibly even viewed as a glorified term for translation. Consequently, should any issues arise with Localized product, it’s clear to these groups where the responsibility lies.

So in response to the initial question I posed, who’s responsible for Localization in your organization? The truth is, in the broadest sense of the term, “Localization”, that everyone at every level of your organization is responsible for Localization (If we take it that by Localization we are in fact referring to GILT).

Just because a Quality or Quality Assurance department may exist within an organization, this does not mean that quality is the sole responsibility of this department and is no longer a concern for the rest of the organization. Similarly Localization, or more accurately Globalization, must be a discrete function of every individual within your organization. If not, there will be an inevitable adverse impact on Internationalization and subsequently the quality of the localized end-product will suffer, as will sales in the target region for that localized product.

Each step within the Globalization, Internationalization, Localization chain will have an exponential impact on the next. If you don’t take your Globalization strategy seriously enough, then, in the absence of a firm mandate from the highest levels of your organization, Internationalization will suffer because there will be no development impetus to properly Internationalize your software. If the Internationalization effort is poor, Localization will be painful, perhaps even impossible within certain software features, and you will be looking at a lengthy delta between your domestic software release and your localized releases.

Conversely, if you start with a solid and coherent Globalization strategy that is communicated, in a relevant and contextual manner, to all levels within the organization, then Internationalization will be an integral part of the SDLC, Localization should be a straightforward, finite task, and you will be in a better position to achieve a Sim-Ship of domestic and localized software releases.

Some people may prefer to use the acronym GILT, some may prefer “glocalization”. For me, the answer to this conundrum, and to addressing people’s sometimes limited awareness of what Localization entails, does not lie in changing terms or the invention of new terms and pseudo-techno-babble. It’s too late. The horse has bolted on that one. It would be comparable to Apple insisting that people stop using “iPod” as a brand name and adopt another title for their pre-existing portable media players. Instead, I believe the answer lies in educating all the relevant stakeholders within an organization on the importance of G11nI18n, and L10n and how these relate to them and various groups throughout the organization in terms of responsibilities.

So with this in mind, in upcoming posts I will take a look at the terms Globalization, Internationalization and Localization in more detail, their inter-dependent relationship, who owns what in terms of responsibilities, what they mean to your organization, and what you should know when endeavouring to sell software in a global marketplace.

Posted in Globalization, Internationalization | Tagged: , , , , , , , , , , , , , | 5 Comments »