On iOS 14, we need to implement AppTrackingTransparency to request the user's permission.
This permission alert always appears in English. I would like to switch to another language, for example Portuguese.
The message I enter in Info.plist appears in Portuguese, but the title and buttons appear in English
The reason is that you are merely using Portuguese words in your app. You have not written and structured an app that is localized for Portuguese. Your app's localization base language is still English.
The system texts, such as the title and buttons of the alert you're talking about, will therefore appear in English even on a device whose language is set as Portuguese.
Related
When I say "Translate English to Spanish" Siri opens the app Google Translate.
How does Siri know to launch this app and is it possible to register your app for certain keywords with Siri?
Usually you can open an app telling Siri something like:
"Open [APP NAME]"
My guess
In your case I guess Siri is not interpreting the sentence. When you say Translate English to Spanish Siri does try to understand what action should take and since no action with an hight Matching Rate is found then Siri tries to open the App having a name equals to the first word of your sentence.
Test 1
In Italian language the app is named Traduttore. When I say Traduci dall'italiano all'inglese (Translate English to Spanish) Siri does NOT open the translator because Traduttore does not match Traduci.
Test 2
If I say: Musica dall'italiano all'inglese (Music italian to english) which is not a real command, Siri does open the Music app.
Siri and third parties apps
Right now the interaction between Siri and third parties apps is pretty limited, however maybe in the future Apple will provide the tools for a deeper interaction.
Maybe something like what is happening with HomeKit where Apple created an interface to make Siri and Home Devices to talk each other.
You should look at Siri's custom vocabulary.
At the moment it is limited to your App's name and Apple specified domains e.g. Workouts or Ride Booking
But if you're lucky enough that your application falls within those categories you can map some more app specific terminology.
For example, if you had a rock climbing application, Ascent, you could start with a voice command similar to "Hey Siri, start climb" rather than "Hey Siri, start workout with Ascent"
All of your localized vocab choices should be placed into their appropriate language-specific project directories (.lproj).
I'm developing iOS application for the blind people. In the beginning of the application user can select his/her preferred language and according to that app's contents(text) will change to selected language.
App support users who are understand English and Tamil.
I don't have any issues if user selected "English" as proffered language, but I need to know how can I convert "Tamil to Speech" voice over in iOS? ( basically a language doesn't support by default in voice over)
Is there another way to achieve this functionality?
I am trying to get some input from the user on the Apple Watch using presentTextInputControllerWithSuggestions. I wonder what happens if user speaks multiple languages – is there a way to detect which language has he spoken?
Also, is there a way to find out what languages are set in his preferences?
Not having a Watch on hand, I don't think anyone here knows. (Edit: this was first posted before the Watch launched.) But even though it'd be really cool if there were dictation software that could guess cual idioma で話しています from word to word, watchOS is no different than iOS in that respect.
In iOS, Siri listens only in the language you set in Settings, and dictation listens only in the language of the active keyboard (whose microphone button you pressed to start dictation).
In watchOS, Siri likewise has a set language. Dictation is based on the keyboard language last used on your paired phone, but you can also change the language during text entry with a force press. That's a user choice for a system service, so it's opaque to the app, just like choice of keyboard is to an iOS app. (You're welcome to perform some sort of text analysis of you want to know what language of text the user has entered.)
I've been tearing my hair out trying to understand how the translation systems works on App Store :
I wanted to release my app to Poland and discovered that polish is not supported on the App Store. Fine ! Let's put some polish references, description and screenshot in my "english" default language while keeping the normal english in the "uk_english" non default language.
(my app is not available in the US).
My app is now live and polish users see the non default "uk-english" description ...
According to : Localized app description does not appear on the App Store I have settled every international parameters of my iPhone in polish, with a polish app store account, and I don't see the default "english" setting, but the non-default "uk english".
Why ?
Typically there is no "default" language. Rather there is a list of languages in order of preference.
Since the UK is closer to Poland than the USA, it makes sense for UK english to be preferred over US english.
I would send an email to Apple asking how to deal with this situation.
In an iPhone app I want the user to do some input in a different language.
For example the user is an English speaker and has his(or her) device set with the English language. When running my app he(or she) will have to input some French and some Japanese.
Is there a way I can bring the proper keyboard from inside the app without forcing the user to change his(or her) usual settings. According to what I have read up to now, I am not sure there is a good answer. I still hope.
Thanks for any tip.
You can usually pick certain keyboards (numeric, email, etc) from the interface-builder-like aspects of UITextFields, so take a peek at the code options for setting those things. You may be able to pick it that way.