I'm attempting to "extract" any data from Adobe tracker. For example, number of comments, reviewer name. I can't find any way to do this so far. I know that Tracker just uses RSS feeds but I can't find WHERE exactly all of this is located and where to pull it from. Has anyone ever done this before? Any help would be appreciated.
I've done it. It's pretty straightforward. The repository where you are collecting the comments will have a bunch of XML files. There'll be a folder for each document at the root of the repository and then an XML (RSS) file for each user in each document folder. Each comment by that user is an "item" in the RSS. There are a bunch of other details about the comment as well and there are extensions to RSS that Adobe uses. Those are in the documentation at the link below.
Tracker Documentation:
http://help.adobe.com/en_US/acrobat/acrobat_dc_sdk/2015/HTMLHelp/Acro12_MasterBook/Tracker_Introduction/About_Acrobat_Tracker.htm
Related
I'm working on a Quizz with Html/JS on Github and which will be dedicated to my comrades.
I would like to be able to read everyone’s answers so I thought about creating a text or csv file with their answers that would be saved in a specific directory of the github project.
But I’m a beginner and I don’t know if that’s possible, i’ve seen tracks that use PHP or NodeJs with FileSaver.js, but I haven’t managed any of them because i would like it to be automatic, not to ask the user to download his answers.
If some people knwo how to do it or explain me why it’s impossible and how to do it otherwise it’ll be cool.
Thanks ! ;)
Unless you want to make every person using the quiz a contributor to your Github project (which will require that they sign up for Github accounts and tell you their account name so you can manually grant them permission) and then use the API to read the CSV file, modify it, then commit the change (and resolve any merge conflicts caused by race conditions): This is not possible (and if you are willing to do that, then it is among the most complex approach that you could take).
If you want to store and aggregate data submitted by visitors to a website then write some server-side code (using whatever language and frameworks you like, PHP and Node.js are both options) and use a web hosting service designed to support them. Github Pages is designed only for static pages and doesn't support any form of server-side programming.
Once you store the data in a file, just use git commands to commit and push it.
I am looking to create a little Google App Script to parse another GAS file to pull out the functions and any jsdoc type comments, and then display it.
Like if you navigate to a script you can get the revision history etc, but I wanted to be able to pass in a Drive folder containing multiple script files and develop a front end to display each's content.
I know about the Google Drive REST API with exportLinks do download as JSON.
I wondered if there was a simpler way to do it from another GAS file, or would I need to URLFetch of some description to do it over the RestAPI?
I have tried searching for similar things but they always lead back to the DriveAPI.
Any help with a starting place, or what GAS Objects I should start looking at?
Thank you
You have to use the REST based Drive API, there is really no other alternative. However, you don't need to use URLFetch to access it. Google Apps Script has what's known as Advance Google Services a collection of interfaces that grant native access to many REST based APIS including the Drive API.
You can check out the Drive specific documentation here. Its somewhat sparse, so you'll often have to fall back on the REST based documentation as well as rely on autocomplete inside the script editor to make sense of things.
I am new to web development;
I am trying to develop some web application to view data and do insert and delete on a little database including two tables.
I use Node.js for my server.
Currently, the data is saved in two XML files. I do my operations(insert,delete,...) by changing the XML at clientside and sending the XML over to server to be saved.
However, I guess this is the hard and native way and there should be some library or some technology to help, because this is such a common scenario.
I also think that using some database rather than XML may help, though the database management utilities are not much needed really.
Note: I don't browse all the data directly from database. the database includes path of some images which are loaded for the user to be viewed.
Any idea is appreciated. Thanks.
I would suggest sqlite as the way to go! U will find an abundance of information on how to use it if U google it.
Have a look at this : https://codeforgeek.com/2014/07/node-sqlite-tutorial/ it may be the answer to the question U are asking :)
Google Drive team has recently announced that third-party Drive app can provide their own thumbnail as a file preview (http://googleappsdeveloper.blogspot.com/2012/10/thumbnails-for-your-custom-file-types.html).
What is for you the simplest way to achieve that?
Our app is using AngularJS on front-end and Python webapp2 on the backend-side.
The fact that the file URL is not public and since we won't make our users' files public, it sounds harder to make it on the server-side so I started my research by looking front-end solution.
I found html2canvas (http://html2canvas.hertzen.com) which sounds great but which does not generate a correct render of our app (missing part).
I was also thinking about something with PhantomJS to generate this preview but it sounds a little bit overkill.
Thanks for your help.
Edit
It actually works, but Google Drive renders it in small (like 300x200px) whereas the original canvas is 630x456px. Any thoughts on this issue?
Please find the original thumbnail and its render in Google Drive attached.
I would have suggested html2canvas, but as you found out it has some limitations which may make it a non-starter. Depending on what you'd like as the preview, a full screenshot might not be necessary. If, for example, you're referring to http://www.videonot.es/, a still/thumbnail image from the video might be a sufficient preview image.
I am making a website for someone and they want to be able to have that site fetch the github readme markdown file at a certain URL and display it on the website, so that instead of having to write the readme in two places, it just pulls from github. Is that possible? How do I do it? I saw this:
https://github.com/coreyti/showdown
which turns markdown into html, but I'm still not how I would fetch the readme URL and convert it into an object that showdown could parse.
Any ideas would be greatly appreciated.
You can use StackEdit. It allows you to publish your markdown document on Github and on others locations at the same time in Markdown or HTML format. For instance, you could publish the HTML on a public Google Drive or Dropbox location.
NOTE: I'm the developer of StackEdit
GitHub has an option to show file source, Raw button in top right corner. Raw link for your example is: https://raw.github.com/coreyti/showdown/master/README.md
Assuming that README file is formatted in Markdown already, you could just fetch source and format it on your side, libraries most probably already exist for your language.
UPDATE
I wouldn't actually download file from GitHub every time the page on your site is requested. GitHub may be down, connection may be slow -- and this will affect visitors of your site. Instead, you may want to have a cron job running on the server that would download a file from GitHub, say, every five minutes, and cache it locally. Then, every time you need to display the file, you'll read a local copy and don't depend on GitHub server being accessible. As a drawback, you will have a certain synchronization delay (5 minutes in my example).
Heroku seems to be doing it by copying the rendered html https://elements.heroku.com/buildpacks/stouffi/heroku-i18n-js-buildpack-ruby#buildpack-instructions
I’ve written riss.awk to insert a README.md on your website, optionally performing some transformations in the process.
be able to have that site fetch the github readme
You can automatically fetch the original README.md file like this:
curl https://raw.githubusercontent.com/cljoly/readme-in-static-site/main/README.md | awk -f riss.awk >readme-in-static-site.md
convert it into an object that showdown could parse
I couldn’t find showdown’s documentation, but there might be some kind of API you could use to upload the file generated above (readme-in-static-site.md)? A cronjob-like task could then run this process every few hours to keep things up to date.