Which way is legal to store Youtbue API data - youtube

Im planning on making a website which stores user playlist data to share with other users in different way than youtube allow (My users will be able to view all records of all the playlists they created in a single list and etc) .
But I wonder if i can store data received from the Youtube API Data (playlist title, description, date, list of videos, thumbs addresses, etc) in my website database or the only way is to store in web server Session.
Of course I know that save of stream is not allowed.
Thank you very much, but I asked them if storing data generated by YouTube API Data (not JSON or XML results but data etc. title, description, id, thumbs addresses etc) in the database is consistent with the rules. The rule is the point:
Caching. Your API Client may employ session-based caching solely of YouTube API results, but You must use commercially reasonable efforts to cause Your API Client to update cached results upon any changes in video metadata. For example, if a video is removed from the YouTube service or made "private" by the video uploader, cached results shall be removed from Your cache. For the avoidance of doubt, Your API Client shall not be designed to cache YouTube audiovisual content.
What did they mean by a "session-based caching" and "YouTube API results".

if you are storing in your database, then you can give additional fields that he one advantage, and allow the user to overwirite the title and description etc, if you don't want the user database then rely on youtube data , these are preotyt XML,JSON data and very lightweight, it' good to implement to save the recenet snapshot in you database

Related

List all Airtable tables via API request

Given that user provided his API key, is there a way to retrieve a list of tables in user's AirTable account?
I do know there is a URL for API queries, with base and table encoded there. Also API docs include only info on how to work with a particular table.
My use case is to allow user to choose a table they want to work with.
Airtable now offers a Metadata Api, however access is limited and must be individually requested, at this time.
As part of our ongoing effort to make Airtable more and more extensible, we’re beginning to expand access to the Metadata API 360.
The metadata API gives you read-only access to a list of a user’s bases, as well as the specific schema of each base - things like what tables exist, what fields exist on those tables, etc.
https://community.airtable.com/t/accessing-the-metadata-api/37080
Metadata Api Documentation
Enterprise API also offers endpoints capable of fetching Metadata.
This would be part of the metadata API which does not exist.
See this community post for some non-api workarounds.

Video name by video ID

I am using the BigQuery Data Transfer Service to get data from YouTube Analytics to BigQuery and from there to Google Data Studio. This works fine, but the tables do not contain the video names. There are just the video IDs.
Best case was if I could get a field 'video_name' in here:
Is there a way to display the video names in Data Studio, e.g. for monitoring "Top Videos By Channel" in a chart?
Here is the youtube's data api, that can serve the purpose of yours,
I tried with this using https://content.googleapis.com/youtube/v3/videos?part=snippet&id= simply you need to provide single or mutiple video_ids like this id=xriwC_UpJER,1BB_dyvjrpf or just single id id=1BB_dyvjrpf.
Keep in mind that you need to be authorized to access this api, i-e you will need access token.

YouTube API derived data

I'm working on a application which would gather YouTube user's video data and create some meaningful data and metrics to help the creators market their videos better and expand their audience.
The problem is that since December 18, if I'm not wrong, this kind of practice is forbidden.
Can someone from Google comment and explain this change? Why can't I create metrics based on YouTube data, even if I visibly communicate that this is not data from YouTube?
For example: I would like to fetch users video description and tell what's the keyword density, how well is it prepared for SEO (in % or something).
And I guess that this new term destroys many businesses which are doing exactly that thing, creating meaningful data based on YouTube API. (Tubular, TubeBuddy, VidIQ).
Please! Anyone?

Import data from another source into Adobe Analytics

I’m trying to tie data from another product with my data inside of Adobe Analytics.
We have Adobe Analytics javascript on our website collecting data and we use a third party tool to track how users interact with certain parts of the website. We’re trying to use the Adobe API to tie the data together.
So far we’ve gone down the path of using the Data Insertion API, but it wasn’t quite right as it’s meant to be used as a replacement for the JS, from what I can tell.
We also explored using the Data Sources API. Now the documentation for this suggests you can use a transaction ID to tie offline data with the data collected from the JS, we’ve tried this and it doesn’t match the data up. We’re now exploring using Visitor ID to tie the sessions together but we’re having problems uploading any rows with the Visitor ID column, Adobe just returns the error “Column header: ‘visitorid’ is not a valid column header”. We’ve tried several different variations of visitor id, such as “visitor_id”, “visitor-id”, “vistor id”, etc and still no luck.
The end goal is for us to be able to upload data to Adobe that will update/add eVars for already existing sessions earlier that day. How would I go about doing this? Is there something I'm missing or doing wrong?
Edit: I managed to solve this problem by using the Adobe SAINT API. When a user arrives at the site, we push an eVar for that user with a unique ID and then the day after we use the SAINT API and the unique ID in the eVar we pushed previously to add the additional data we needed.
It could be a good idea to look back at the Data Insertion API and combine it with the visitorId approach where you tie existing/old visitorID's to new eVars and use the timestamp to "update" the dataset.
Although this is experimental, it might be worth a try.
Best regards,

Getting curated content using streaming API

I cam across one third party API which provides organized twitter data as per user request.
example
http://api.frrole.com/v1/curated-content?location=India&contenttype=link&category=sports&orderby=popularity&minrt=10&resultcount=100&apikey=4534sdg34343
It give top 100 tweets from india, which contains link in tweets and falling into sports category and having minimum rt count =10.
Can someone tell me how does this work? It gives json data as result. Does this result comes directly via twitter streaming APIs or third party's own data base?
Is there any way for streaming API to do this?
This is likely coming from their database. They are consuming the data from the Streaming API or another API and then applying their categorization to it, then giving you the ability to access it from an API.
Some of the criteria above you could get from the Streaming API such as the filter_level (to find top Tweets) but you wouldn't be able to get that whole set of requirements (especially the sports categorization).

Resources