Emoji support in Google's Cloud Speech API - google-cloud-speech

I've noticed that certain apps on Android (ie. gboard) support translating phrases such as 'poop emoji' into the actual emoji as part of speech recognition. I was wondering if this is something that is supported through google's cloud speech APIs that I could similarly use in my own applications?
In my initial scan of the API I can't see anything that might indicate a way to turn this on (ie. RecognitionConfig et.al has no obvious toggles for it), and in some quick one-off tests in my own app I wasn't provided emoji-fied results from the service.
I've done a bunch of googling but found nothing so far.
Any insight here would be awesome, thanks!
-edit- Thanks to the answer below I have learned this currently is not supported. I've gone to Google's issue tracker to request this feature. If anyone wishes to track the feature request the link is:
https://issuetracker.google.com/u/1/issues/113978818

The Cloud Speech-to-Text API service doesn't currently support emoji phrases recognition; however, you can use the Send Feedback button located at the lower left and upper right corners of the service public documentation, as well as take a look the Issue Tracker tool in case you want to raise a Speech API feature request in order to notify to Google about this desired functionality.
Finally, you can refer to the Release Notes section of Speech-to-Text API to keep the track of the new features and functionalities added to the service.

Related

How to use Google Tag Manager for iOS (swift) with GCP Recommendations AI?

We are planning to make a recommendations engine for the features in the app and we were thinking about using GCP Recommendations AI. Then, we ran into some unclear areas due to lack of examples.
Can we use Google Tag Manager for iOS (with Firebase Analytics) to send real-time event data to GCP Recommendations AI? Could you please provide some docs/examples/articles that explains how (We are using Swift to code the app)? It is possible for web sites, according to the docs.
You can not. Unless it's not a native app.
If it is a web-app, then sure you can use the power of JS to send whatever you need to any endpoint. We use Segment for real time personalization, but even that requires quite a lot of involvement from the app teams. And no, GTM is not helpful for native apps.
If you already have custom analytics tracking implemented in your app, then there's a high chance that the devs used a wrapper around the firebase/gtm sdk to send events in a more comfortable way or add adequate logging. In that wrapper then, the devs could implement the event streaming to your GCP recommendations AI in case it has the capability to store data.
If it does not, you can still set up a neat app engine (or a differnet endpoint) and stream data to it. The app engine would allow you to tweak, sanitize and debug the data, after which you can stream them to something like Big Query, which your Recommendation AI would likely to be happy to work with.
I'm talking about app engine and big query because there's a nice repo I've used a few years ago... Here: https://github.com/MarkEdmondson1234/ga-bq-stream
This repo is attempting to use front-end GTM to stream events, but you won't be able to use it (unless you're dealing with a web app). So instead, you stream them in a very similar way from your analytics wrapper.

What is best approach making Google Home Realtime speaking

I have an idea of making a live commentary assistant for football matches and so far this is what I have achieved:
I am using Dialogflow and achieved linking it with Actions on Google, so every time I say something Dialogflow can detect and fire an event to google home so the Google home responses.
The thing I would like to know more is, how about for example when a team scores a goal and google assistant catch that instantly, in real time.
One possibility is, I have the API for getting all the matches and scores updated every time team scores and fetch that every second and then call an intent in dialogflow so that will fire to google home assistant, but I am thinking that is somehow not the best idea.
Does anyone have any idea about this?
Sorry I am not adding any code, as the code in this case is not important as is the approach, idea how to achieve it.
Unfortunately, Actions on Google is not suited for this kind of use-case. The platform is designed for conversational experiences, where there's a back and forth. The platform intentionally limits real time features like background continual listening as well as things like push notifications.
Push notifications do work on phones, although not other surfaces like smart speakers (ie. Google Home). You can use that to get close to the behavior you may want, but otherwise it may not be suitable for your use case.

Using google assistant SDK with raw text as user input

I am currently working on a project that uses the Google Assistant SDK with Python. I have it working with direct audio listening, but I want to know if there is a way to use it with raw text input instead of listening to audio.
This is, apparently, a common request - but there is no way to do it yet. (Given this is still an early Developer Preview, and there have been many requests for this, we can hope they'll deliver it as part of a forthcoming update.)

UBER api, can I retrieve driver fare for each ride (for a driver's utility)?

I am building a small utility for uber drivers. Currently when they drop off a passenger and eventually get the resulting fare processed and returned to them, I have them manually enter it into the utility. Is there a way to 'catch' this information programmatically and automatically populate my utility when this becomes available to the partner application?
We don't currently make this kind of data generally available via the API. But thanks for your interest and the question, it helps us to understand what the community is looking for so we can prioritize what gets built and released next.
Keep an eye on the Uber Developers blog for news around upcoming APIs.

Speech recognition from recorded file

I've been researching several iOS speech recognition frameworks and have found it hard to accomplish something I would think is pretty straightforward.
I have an app that allows people to record their voices. After a recording is made, they have the option to create a text version.
Looking into the services out there (i.e., Nuance) most require you to use the microphone. OpenEars allows you to do this, but the dictionary is so limited because it is an offline solution (they recommend 300 or less words).
There are a few other things going on with the app that would make it very unappealing to switch from the current recording method. For what it is worth, I am using the Amazing Audio Engine framework.
Anyone have any other suggestions for frameworks. Or is there a way to dig deeper with Nuance to transcribe a recorded file?
Thank you for your time.
For services, there are a few cloud based hosted speech recognition services you can use. You simply post the audio file to their URL and receive back the text. Most of them don't have any constraint on the vocabulary. You can of course choose any recording method you like.
See here: Server-side Voice Recognition . Many of them offer free trial as well.

Resources